Message splitting-- Queue stopped when no message created for one interface

Hi all,
I think you saw one post regarding this issue in SDN, but I can't find the post any more.
So, I have message splitting (it's multiple mappings to map one outbound msg to multip inbound msgs, not one mapping to map 1 to n. This is because of our own reason, but should not affect the analysis of the issue) to split one message to multiple. It works fine if the outbound message contains data for every inbound interface. However if the outbound message does NOT contain data for any one inbound message, there is an error in MONI and all subsequently splitted messages will not be passed in IE. The error in MONI says Split mapping created no messages, and Queue status turns to be Queue Stopped, and thus all the subsequently splitted messages are stuck in the queue.
Why this? and how to solve this problem?
Thanks
Jayson

>
Jayson wrote:
> Hi all,
>
> I think you saw one post regarding this issue in SDN, but I can't find the post any more.
>
> So, I have message splitting (it's multiple mappings to map one outbound msg to multip inbound msgs, not one mapping to map 1 to n. This is because of our own reason, but should not affect the analysis of the issue) to split one message to multiple. It works fine if the outbound message contains data for every inbound interface. However if the outbound message does NOT contain data for any one inbound message, there is an error in MONI and all subsequently splitted messages will not be passed in IE. The error in MONI says Split mapping created no messages, and Queue status turns to be Queue Stopped, and thus all the subsequently splitted messages are stuck in the queue.
>
> Why this? and how to solve this problem?
>
> Thanks
> Jayson
Hi Jyson,
what i analyse from this is you may be using different maapings and different inbound interfaces
but your receiver is the same and hence you are using only one interface determination with all the   interface mappings specified there...
if this is the case..i suggest you remove the "maintain order at runtime "
tick in that interface detemination and it will surely work
giving points is another way to say thanks
Edited by: Tarang Shah on Mar 20, 2009 7:18 PM

Similar Messages

  • Logical standby stopped when trying to create partitions on primary(Urgent

    RDBMS Version: 10.2.0.3
    Operating System and Version: Solaris 5.9
    Error Number (if applicable): ORA-1119
    Product (i.e. SQL*Loader, Import, etc.): Data Guard on RAC
    Product Version: 10.2.0.3
    logical standby stopped when trying to create partitions on primary(Urgent)
    Primary is a 2node RAC ON ASM, we implemented partitions on primar.
    Logical standby stopped appling logs.
    Below is the alert.log for logical stdby:
    Current log# 4 seq# 860 mem# 0: +RT06_DATA/rt06/onlinelog/group_4.477.635601281
    Current log# 4 seq# 860 mem# 1: +RECO/rt06/onlinelog/group_4.280.635601287
    Fri Oct 19 10:41:34 2007
    create tablespace INVACC200740 logging datafile '+OT06_DATA' size 10M AUTOEXTEND ON NEXT 5M MAXSIZE 1000M EXTENT MANAGEMENT LOCAL
    Fri Oct 19 10:41:34 2007
    ORA-1119 signalled during: create tablespace INVACC200740 logging datafile '+OT06_DATA' size 10M AUTOEXTEND ON NEXT 5M MAXSIZE 1000M EXTENT MANAGEMENT LOCAL...
    LOGSTDBY status: ORA-01119: error in creating database file '+OT06_DATA'
    ORA-17502: ksfdcre:4 Failed to create file +OT06_DATA
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    LOGSTDBY Apply process P004 pid=49 OS id=16403 stopped
    Fri Oct 19 10:41:34 2007
    Errors in file /u01/app/oracle/admin/RT06/bdump/rt06_lsp0_16387.trc:
    ORA-12801: error signaled in parallel query server P004
    ORA-01119: error in creating database file '+OT06_DATA'
    ORA-17502: ksfdcre:4 Failed to create file +OT06_DATA
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    ORA-15001: diskgroup "OT06_DATA" does not exist or
    Here is the trace file info:
    /u01/app/oracle/admin/RT06/bdump/rt06_lsp0_16387.trc
    Oracle Database 10g Enterprise Edition Release 10.2.0.3.0 - 64bit Production
    With the Partitioning, OLAP and Data Mining options
    ORACLE_HOME = /u01/app/oracle/product/10.2.0
    System name: SunOS
    Node name: iscsv341.newbreed.com
    Release: 5.9
    Version: Generic_118558-28
    Machine: sun4u
    Instance name: RT06
    Redo thread mounted by this instance: 1
    Oracle process number: 16
    Unix process pid: 16387, image: [email protected] (LSP0)
    *** 2007-10-19 10:41:34.804
    *** SERVICE NAME:(SYS$BACKGROUND) 2007-10-19 10:41:34.802
    *** SESSION ID:(1614.205) 2007-10-19 10:41:34.802
    knahcapplymain: encountered error=12801
    *** 2007-10-19 10:41:34.804
    ksedmp: internal or fatal error
    ORA-12801: error signaled in parallel query server P004
    ORA-01119: error in creating database file '+OT06_DATA'
    ORA-17502: ksfdcre:4 Failed to create file +OT06_DATA
    ORA-15001: diskgroup "OT06_DATA" does not exist or is not mounted
    ORA-15001: diskgroup "OT06_DATA" does not exist or
    KNACDMP: *******************************************************
    KNACDMP: Dumping apply coordinator's context at 7fffd9e8
    KNACDMP: Apply Engine # 0
    KNACDMP: Apply Engine name
    KNACDMP: Coordinator's Watermarks ------------------------------
    KNACDMP: Apply High Watermark = 0x0000.0132b0bc
    Sorry our primary database file structure is different from stdby, we used db_file_name_convert in the init.ora, it look like this:
    *.db_file_multiblock_read_count=16
    *.db_file_name_convert='+OT06_DATA/OT06TSG001/','+RT06_DATA/RT06/','+RECO/OT06TSG001','+RECO/RT06'
    *.db_files=2000
    *.db_name='OT06'
    *.db_recovery_file_dest='+RECO'
    Is there any thing wrong in this parameter.
    I tried this parameter before for cloning using rman backup. This din't work.
    What exactly must be done? for db_file_name_convert to work.
    Even in this case i think this is the problem its not converting the location and the logical halts.
    help me out.....
    let me know if you have any questions.
    Thanks Regards
    Raghavendra rao Yella.

    Hi reega,
    Thanks for your reply, our logical stdby has '+RT06_DATA/RT06'
    and primary has '+OT06_DATA/OT06TSG001'
    so we are using db_file_name_convert init parameter but it doesn't work.
    Is there any thing particular steps hiding to use this parameter? as i tried this parameter for rman cloning it din't work, as a workaround i used rman set new name command for clonning.
    Let me know if you have any questions.
    Thanks in advance.

  • Two idocs are created for one output message type

    hi all,
    we are communicating our sap idocs to external system using ALE.It is working smoothly.
    Our problem raises here,
    .idocs are creating at the time of output type attachment for purchase orders.But rarely,two idocs are creating for one message type.It means two idocs are created for same Purchase order.It makes complications for the external system.
    anyone can help me pls?
    thnks in advance..........

    thanks jurgen for ur reply,
    jurgen, the second idoc is not an "ORDCHG",It is the replica of first idoc.This one has also "ORDERS" message type.The only difference between these idocs are the time.It means the difference between the time field segments of these idocs have only three(3)seconds difference.We think that its from the system error.If it is not from the system error pls give me the explanation.
    thanks in advance.......

  • My form has a list of yes or no questions.  I want to use either cheboxes or radio buttons.  I can't make it work.  When I check yes for one question, it marks all yes answers for all other questions.  If I mark some radio buttons to answer, it unmarks la

    My form has a list of yes or no questions.  I want to use either cheboxes or radio buttons.  Unfortunatly I can't make it work.  When I check yes for one question, it marks all yes answers for all other questions.  If I mark some radio buttons to answer, it unmarks later.  Need immediate help!!!

    Is it that you are using a PDF-based form? Did you copy and paste the Yes/No fields all the way down your form? If so, then all of the yes/no boxes are copies of each other and have the same name and properties in the PDF.
    For example, if you had:
         Question 1 and Question 1 Yes/No checkboxes
         Question 2 and Question 1 Yes/No checkboxes
         Question 3 and Question 1 Yes/No checkboxes
         Question 4 and Question 1 Yes/No checkboxes
    Then whatever answer you selected in Question 1 would populate down through the rest of the form.
    Changing the Yes/No field properties in questions two through four would eliminate the problem.
    I hope that helps,
    Brian

  • Queues Stopped when Dispatcher Fails

    Hi ,
    We had an issue with our Web dispatcher and most of the messages failed in the step "Receiver Grouping" with the error "HTTP_CLIENT_RECEIVE_FAILED" when calling the mapping runtime.
    After sometime we also noticed the queues failing with the error "CLient receive failed:Queues stopped".
    I am just wondering what the role of the web dispatcher could be in calling the mapping runtime.Can somebody please explain this to me ?
    When a message is already being processed on an instance on the Int. Engine, the IE will call the JAVA stack on the same instance(if I am not wrong) then how can the error occur if the Web dispatcher is down.Moreover, in our visual admin properties we have checked the JCA Adapter settings and it is calling its "LOCALHOST" always.
    Please explain me in detail the mapping runtime step.
    Thanks,
    Harish

    hi,
    I'm not a basis consultant... but I know you can use a web dispatcher, for instance when in production you have several PI application servers. by this way, the IE can communicate with the AE by using just one address (your web dispatcher).
    So if this last one shut down, your connexion IE -> AE works no more, and your messages are in error in PI queues (SMQ2.)
    regards.
    Mickael

  • Test message from RWB: Error when sending message: 401 Unauthorized

    Hi Experts,
    While I am sending test message from RWB, I am getting following error:
    Error when sending message: 401 Unauthorized
    I am using correct userid/password as I am working with Repository/directory. Also in one of the test scenario files are picking up from folder but never got processed because of similar issue (HTTP_RESPONSE_CODE_NOT_OK) and SERVICE NOT AVAILABLE.
    I am having SAP_ALL authorizations and few days back everything was fine.
    I think something wrong with configuration in XI but I am unable to pin point.
    Please help.
    Regards,
    Gourav

    Yes,
    but
    Role of Business System : Integration system
    Corresponding Integ. Server : http://<XI server>:<http port>/sap/xi/engine?type=entry
    How can I change it to Application system
    and RUNTIME->IS_URL = http://<XI server>:<http port>/sap/xi/engine?type=entry
    Regards,
    Gourav

  • Number of unread message is in negative when all message are read

    Hi, in my list of text message of my iphone, when I have all the message read, the number of unread message is in negative!
    Thank you for your solutions!!

    have you taken the basic troubleshooting steps outlined in the User Guide?  What exactly HAVE you done to try and solve your problem?

  • Elements 10 organizer keeps stopping when trying to create a slide program.

    When trying to create a slide program Elements 10 Organizers stops working about half way through the selection.  How can I correct this problem?

    Thanks Lennart for your help...you are the man!
    I tried what you said and was able to import my movie into iDVD although I still had trouble writing the DVD (4 more times) as the titles page viewed with the wrong buttons and the sound was out of synch with the movie.
    I tried once again from within iMovie and clicked the 'proceed' button instead of 'render and proceed' and this seemed to work okay.
    I then went on to create the DVD which worked fantastic.
    Thanks for your help which lead me in the right direction.
    Julie

  • Pen pressure sensitivity stops when I have more than one window open

    I recently upgraded to Flash CC and have noticed some new bugs. Here's one of them:
    When I have more than one document open at a time, my pen pressure sensitivity stops working. In fact, it seems to work on the newest opened document, but will stop working on any previosuly opened documents. If all but one document is closed, the pen pressure returns.
    I'm using Windows 8, and a Cintiq 21UX.
    Steps to reproduce:
    - Launch Flash.
    - Open a new document.
    - Select the "Brush" tool.
    - Toggle "Use pressure" in the tool bar.
    - Draw something to make sure the pen pressure is working.
    - Open a new document.
    - Draw something to make sure the pen pressure is working.
    - Return to the previous document and attempt to draw.
    - Notice how the pen pressure is absent.
    Any help or suggestions would be great. Thank you.

    No problem.  I've been using older versions of Flash until CC gets the bugs squashed.  This pressure issue is a deal breaker for my coworkers and me, and also this audio one: http://forums.adobe.com/message/6013488?tstart=0

  • EDI message is not sending to the third party for one particular SA

    DearAll,
                   We have the scenario to send the schedule lines to the third party system. for one particular schedule agreement schedule lines not sending to the third party system.
    while processing the messgae it is giving the error as "Message XEDI already processed on 20110317 070042"
    What could be the reason?.. we have checked all the EDI settings maintained for this supplier. some two weeks back there is no problem now only getting this message.
    how to resolve this?..
    Thanks in Advance
    Ananth.

    Hi Hareesh,
    Thanks a lot for the help. I have tried with removing the xml namespace, still the same result. I believe, the problem is with MTs tag in response message structure, bank xml data is trying to map with the response structure in PI, but when its finding the MTs tag I believe then its not mapping, and directly its passing the response in SXMB_MONI without any mapping.
    My question is how can I define my response structure without the MTs tagm, I mean same as what I am receiving from bank.
    Also I tried with without any mapping in response structure, I mean I used only one Message type for "response from bank" to "response to ECC" , without xml namespace, in this case I am able to receive the response in ECC, but there is exception called "PARSE_APPLICATION_DATA Error during XML => ABAP
    conversion: Response Message; CX_ST_MATCH_ELEMENT in " while receiving response, and when I go and see the logs, it says "System expected element 'MT_Response_MT940_Test'" , as its try to match with the defined structure in PI.
    Please experts help me, so my question is how can I avoid the MTs tag in my response structure, I just want to define my response structure in PI same as bank is sending me the data in xml tag, without any MTs tag.
    Thanks,
    Farhan

  • Split tunnel works... but only for one IP

    Hi All,
    Dealing with a really frustrating problem. Our setup, roughly, is as follows:
    - We have a remote access VPN that users connect to with any connect; in turn, they are assigned a local LAN address: 10.1.11.192-10.1.11.200
    - We have a site-to-site VPN that connects to Amazon AWS to access 10.0.249.0 and other subnets, and now certain hosts on the Amazon *public* network (e.g, 54.1.2.3). This is done via a split tunnel.
    What we're seeing is this:
    - Users connect to the VPN and are assigned one of the addresses above. Let's use 10.1.11.192 for this example.
    - They can then access anything in the 10.0.249.0 subnet (via the split tunnel) just fine. This goes across both ASA devices.
    - They can then access anything in the public Amazon network (via the split tunnel) just fine. This should only use the remote access ASA.
    So, it seemed like everything was working. When connected to the VPN, Amazon hosts in both 10.x.x.x networks and public IPs that I had specifically tunneled (we plan to transition these to a VPC soon) were accessible, and access happened via the remote access VPN IP (i.e, when connecting to 54.1.2.3, it showed the user being logged in from the Cisco's gateway's IP address, as opposed to the local client IP).
    Now, here's where things got weird: *public* tunneled hosts at Amazon only works with the first address in the pool, 10.1.11.192. No other addresses work. 10.0.249.x is always available, regardless of assigned IP. 54.x.y.z is only available with .192.
    I've used the same computer with different assigned IPs (10.1.11.193-10.1.11.200), and none work. I've connected using different computers.. they work if assigned .192, but not any other addresses. Other users report the same issue.
    TCP handshaking is failing
    I'll use our IRC server (and sometimes ssh server) for testing. I can see my client laptop with a SYN_SENT on that specific connection. I can see the IRC server with a SYN_RECV, and the ASA shows a SYN timeout after 30 seconds. Thus, it looks as though packets from the IRC server can't make their way back through the ASA to my client laptop.
    I suspect this has something to do with dynamic vs. static NAT, etc., but I've fiddled with every setting I can, and coming up blank.
    I'm also baffled as to why .192 works, but no other addresses do.
    I've attached our configuration, minus keys and passwords and certain IPs/hostnames. It's a little bit ugly because it has some poor attempts at fixing this, things I'll probably remove after it works, but.. Could it have anything to do with TCP sequence randomization?
    Thank you in advance for any help.

    Hi Jouni! Thank you so much for your quick reply. Mine was delayed because I've continued to fight the ASA this afternoon! Apologies for my verbosity, always found it better to include too much info than too little!
    To clarify, ideally, a VPN client that connects to this ASA (10.1.11.5) should have three possible destinations:
    - The internet at large - this should not go through the ASA, but should go through the client's own gateway, bypassing this VPN entirely. This works.
    - My servers in an Amazon VPC (10.0.249.x and some others) - this should come in to this ASA (10.1.11.5), which in turn is already successfully routed to another ASA (10.1.11.4 in the config). This works, too.
    - My servers at Amazon AWS that are on the public internet (example below: 107.22.xxx.yyy). This should only involve the one ASA in question (10.1.11.5). This is where I'm running into issues.
    The split tunnel, thus, includes the networks 10.0.249.x and specific public AWS hosts like 107.22.xxx.yyy, etc. I want 10.0.249.x to go to Amazon via the other ASA 10.1.11.4 (again, this works) and I want 107.22.xxx.yyy, etc. to go to Amazon via *this* ASA (10.1.11.5). Basically, it's this last part that's causing problems: it works only one in very, very narrow situation: when the client is assigned the address 10.1.11.192. If a user logs in and is assigned any other address, they won't be able to access the "public" AWS servers.
    I began by taking your advice and moving the IP assignment to an entirely different subnet: 10.1.12.100-10.1.12.200. I then spent a decent chunk of the afternoon adjusting NAT rules (and removing ACLs, per your suggestion). The only brief success I had was when I had about a bajillion NAT rules, and somehow I made it on to the IRC server! Oddly, my connection used the ASA's address itself, rather than the ASA gateway's address. Unfortunately, I stopped being able to reach 10.0.249.x at the time :/ I tried to adjust further, but cannot restore/recall what the NAT rules were.
    I'm miffed because 10.1.11.192 works, and *nothing* else does. There is nothing special about this address in the config - I've tried everything I can to figure out why this one address is different. I've also looked in places that *shouldn't* matter because this bypasses the VPC entirely (AWS security groups, AWS iptables rules, VPC options, etc.)
    Again, the behavior of 10.1.11.193-200 is a partial TCP handshake: the local computer sees SYN_SENT, and the remote server has SYN_RECV, but no connection is ultimately established. This is why I think it could be a NAT issue, but I'm obviously a bit lost here Here's a sample connection attempt, with DNS traffic removed:
    2013-10-17 18:29:09.100 [DEBUG] Message from Host: 10.1.11.5 - Message - %ASA-6-302013: Built inbound TCP connection 43606 for outside:10.1.11.193/56626 (10.1.11.193/56626)(LOCAL\justinsTestMac) to outside:107.22.xxx.yyy/6667 (107.22.xxx.yyy/6667) (justinsTestMac) (pid:25912)2013-10-17 18:29:39.129 [DEBUG] Message from Host: 10.1.11.5 - Message - %ASA-6-302014: Teardown TCP connection 43606 for outside:10.1.11.193/56626(LOCAL\justinsTestMac) to outside:107.22.xxx.yyy/6667 duration 0:00:30 bytes 0 SYN Timeout (justinsTestMac) (pid:25912)2013-10-17 18:29:45.293 [DEBUG] Message from Host: 10.1.11.5 - Message - %ASA-6-302013: Built inbound TCP connection 43612 for outside:10.1.11.193/56626 (10.1.11.193/56626)(LOCAL\justinsTestMac) to outside:107.22.xxx.yyy/6667 (107.22.xxx.yyy/6667) (justinsTestMac) (pid:25912)102 (justinsTestMac) (pid:25912)2013-10-17 18:30:15.322 [DEBUG] Message from Host: 10.1.11.5 - Message - %ASA-6-302014: Teardown TCP connection 43612 for outside:10.1.11.193/56626(LOCAL\justinsTestMac) to outside:107.22.xxx.yyy/6667 duration 0:00:30 bytes 0 SYN Timeout (justinsTestMac) (pid:25912)2013-10-17 18:30:17.976 [DEBUG] Message from Host: 10.1.11.5 - Message - %ASA-6-302013: Built inbound TCP connection 43617 for outside:10.1.11.193/56626 (10.1.11.193/56626)(LOCAL\justinsTestMac) to outside:107.22.xxx.yyy/6667 (107.22.xxx.yyy/6667) (justinsTestMac) (pid:25912)2013-10-17 18:30:48.400 [DEBUG] Message from Host: 10.1.11.5 - Message - %ASA-6-302014: Teardown TCP connection 43617 for outside:10.1.11.193/56626(LOCAL\justinsTestMac) to outside:107.22.xxx.yyy/6667 duration 0:00:30 bytes 0 SYN Timeout (justinsTestMac) (pid:25912)
    Thanks for any insight you have!

  • When we  will go for Ectended Interface Determination

    extended interface determination example give me any one...
    Regards,
    Vinay

    Enhanced (Mapping-Based) Interface Determination 
    Use
    In an enhanced interface determination, you do not enter the inbound interfaces manually, but first select a multi-mapping. You get the inbound interfaces from the target interfaces of the multi-mapping. The inbound interfaces are determined at runtime during the mapping step.
    You typically use an enhanced interface determination if the source message has an element with occurrence 0 ... unbounded (for multiple items of a data record) and you want multiple messages (for the individual items) to be generated at runtime.
    For example, you want to split an overall booking for a trip comprising connecting flights into individual booking orders (for each leg of the trip).
    You cannot realize this usage case with a standard interface determination (without using an integration process). A standard interface determination does allow you to specify multiple (N) inbound interfaces (with different mapping). In this case, N messages with the same payload and different receiver interfaces are generated from the source message before the mapping step; these are then transformed differently depending on the receiver interface. Therefore it is conceivable to write the mappings in such a way that the first mapping from the source message generates a message that only contains the first item, the second mapping generates a messages that only contains the second item, and so on. However, this no longer works if the number of items can vary with each new source message.
    You can use an enhanced interface determination to configure a mapping-based message split for this usage case. You assign the interface determination a multi-mapping that has a target interface with an element with occurrence 0...unbounded. At runtime, the individual messages (for the individual items) are calculated in the mapping step. First, the individual messages are grouped into a bulk message. Then, the bulk message is transferred to the Adapter Engine. The Adapter Engine then splits the bulk message up into the individual messages (see figure).
    Procedure at Runtime During Message Split
    Note that both the bulk message and all individual messages each have a message header with a receiver interface (see figure).
    Receiver Interfaces of Bulk Messages and Individual Messages
    The header of the individual messages contains the relevant receiver interface. It is based on the definition of the multi-mapping. Note that the receiver interfaces of the individual messages may be different. The receiver interface of the bulk message is always InterfaceCollection (namespace http://sap.com/xi/XI/System).
    If the message split only results in one message, the original message structure remains the same. In this case, a bulk message containing just one individual message is not created.
    You can also assign a multi-mapping with multiple different target interfaces to an enhanced interface determination (1:n transformation). Each target interface can contain elements with occurrence 0 ... unbounded.
    Messages cannot be sent by using different Adapter Engines in a mapping-based message split. This affects configuration thus: All receiver agreements that have a receiver interface from the mapping entered in the key must only be assigned communication channels with the following adapter types:
    -          RFC adapter
    -          SAP Business Connector adapter
    -          File/FTP adapter
    -          JDBC adapter
    -          JMS adapter
    -          SOAP adapter
    -          Marketplace adapter
    -          Mail adapter
    -          RNIF adapter
    -          CIDX adapter
    The adapters also have to all run on the same Adapter Engine.
    Adapters developed by partners also support a mapping-based message split if they run on the same Adapter Engine.
    Attachments from the original message are not appended to the messages resulting from the message split.
    Activities
    To execute a mapping-based message split, perform the following steps:
           1.      Integration Repository: Define the multi-mapping (see Developing Multi-Mappings for Message Splits).
           2.      Integration Directory: Define the interface determination.
    Note the following:
    ○     The outbound interface of the multi-mapping must be entered as the sender interface in the key of the interface determination.
    ○     The interface determination type must be set to Enhanced.
    ○     Select the multi-mapping you defined previously from the Integration Repository.
    To select the interface mapping, use the input help ( ).
    The target interfaces of the interface mapping are displayed in the Inbound Interfaces frame.
    The number of messages (for an inbound interface) created in the mapping step is displayed in the Occurrence column.
    regards,
    Raj

  • Multiple Posting change notice created for one material document

    Hi All,
    The scenario is that the user needs to transfer material from unrestricted to quality stock type. while doing so the storage type is different than the actual storage type for the material. Meaning if the material is normally stored in high rack storage type, when it is found that the material is damaged or expired it need to be moved to the quality storage type. For this the user will first create  material document in IM with movement type 322 transferring material from unrestricted stock to qualtiy stock. A PCN will be created in back ground which will in turn be converted manually by the user to a TO. while doing so the user will pick the quant manually from the required storage type and select the destination storage type as the quality or quanranntine area.
    The issue here is that system creates one PCN for each material document line item. The requirement is that the system should create only one PCN with multiple lines for one Material document.
    The setting in the movememnt type clearly says that it will behave as it is behaving.
    Is there any other way or setting by which we can create one PCN with multiple line items for a material document.
    Would look forward to your replies.
    Regards,
    Jeetendra

    Thanks and Sorry for the delayed reply. We have now convinced the users that the system behaviour is standard and we cannot change the same.
    To your question about the need to have one PCN with multiple lines please note that the setup is such that when material document is created for transferring material from unrestricted to quality and vice versa this document is created with multiple materials in one go. Now when the PCN is to be converted to TO then the user ends up in creating number of TO's equal to the number of lines in the Material Document. We wanted to avoid this and hence this was posted in SDN.

  • Two different invoices created for  one order for two different lines

    Hi All,
    I have an order with four line items that was shipped on the same day, invoiced on the same day and all lines have the same delivery number yet two inovices were created for this order..one for the first three lines and one for the last line.
    Can anyone please explain what can cause this to happen?
    Thanks
    A/A

    Hi Asgar,
    If a single order has been split into two invoices, you should check the values of the mandatory columns. During Autoinvoice, the grouping is based on the madatory columns and also your grouping rules. If any of the mandatory column has a different value, it will split into different invoice.
    Please refer the following metalink note for further details
    Subject: How Autoinvoice Grouping Works
    Doc ID: 98772.1
    Hope that helps.
    Thanks,
    Anil

  • Auto  filter settings taking when executed the query for one user id......

    Hi All,
    We are facing one strange issue for  one particular user id  ( i.e for top management) ,when this user executes any of the report  by in putting his own selection parameters say date 15.09.2009 and plant 1111 the output of the report is coming as 15.06.2009 to 19.06.2009 and the plant is also coming as 2222 instead of 1111.
    I have checked it there are variants for this user id .
    When i checked in  the filters in portal it is showing for the cal day15.06.2009 to 19.06.2009,plant 2222,Materials Xyz,Keyfigures say Net amount only, nd more over its getting filtered onalso..... something there is a problem its taking automatic filter values.
    What could be the issue does any one faced this kind of problem?
    Pl help me out by providing your valuable inputs.
    Rgds,
    Rakesh

    closed

Maybe you are looking for