FND_FILE not processing

FND_FILE is not generating the messages in log files.
It generated the custom messages in log files for some days and it didnot generated the messages in log files.
The problem is in happening in PROD instance and It is perfectly working in Test instance
Could you please help me what could be the possible reason?
Thanks

The values of APPLPTMP, APPLTMP, UTL_FILE_DIR where in One of our test instance FND_FILE was writing custom messages in log file perfectly are as follows
SQL> show parameter utl_file_dir
NAME TYPE VALUE
utl_file_dir string /lpar2_stage/SNBX/utl, /lpar2_
stage/XXAFP, /lpar2_stage/SNBX
/inbound, /lpar2_stage/SNBX/in
bound/edi, /lpar2_stage/SNBX/o
utbound, /lpar2_stage/SNBX/out
bound/edi, /lpar2_stage/SNBX/o
utbound/ct, /lpar2_stage/SNBX/
fax, /lpar2_stage/SNBX/inbound
/edi/EEP, /lpar2_stage/SNBX/ou
tbound/edi/EEP, /lpar2_stage/S
NBX/outbound/ACH, /lpar2_stage
NAME TYPE VALUE
/SNBX/inbound/ACH, /lpar2_stag
e/SNBX/outbound/ACH/ARCHIVE, /
lpar2_stage/SNBX/outbound/POSI
TIVE_PAY, /lpar2_stage/SNBX/ou
tbound/WIRE, /lpar2_stage/SNBX
/outbound/POSITIVE_PAY/Archive
, /lpar2_app/snbx/snbxcomn/adm
in/log/SNBX_lin60app2, /snbx/s
nbxdb/11g/11.2.0/appsutil/outb
ound/SNBX_lpar2, /snbx/snbxdb/
11g/11.2.0/appsutil/outbound/S
NAME TYPE VALUE
NBX_lpar2, /snbx/snbxdb/11g/11
.2.0/appsutil/outbound/SNBX_lp
ar2, /usr/tmp
$ echo $APPLPTMP VALUE IS "/lpar2_stage/SNBX/utl"
$ echo $APPLTMP VALUE IS "/lpar2_app/snbx/snbxcomn/temp"

Similar Messages

  • Job getting cancelled for 'Chain not processed for date-2007031920070318' ?

    Hi All,
      We have 2 process chains running only on weekdays - one for sales and the other for billing. A cancelled job is created for one of the process chains with the job log message as 'Chain not processed for date-2007031920070318' for the not run days whereas the other does not have any issues though the same monitor job program runs for both. The difference is only in the variant but nothing much other than that its maintained for 2 different chains.
       We also debugged the monitor job program. It basically does nothing other than displaying messages depending on the status of the chain log.
    Please let me know your suggestions as to why it happens.
    Thanks,
    krishna

    it is also giving following error message:
    Caution: Program has changed
    Caution: At time of termination,  Active source code no longer available  and after above two statements its giving sign "?" before every statement in below manner:
    024300 ?     lit_ekpo_dummy1[] = it_ekpo[].                     
    024310 ?     DELETE lit_ekpo_dummy1 WHERE ebeln NE tab1-ebeln.  
    024320 ?     DESCRIBE TABLE lit_ekpo_dummy1 LINES count01.      
    024330 ?                                                        
    024340 ?     REFRESH lit_ematn_dummy[].                         
    024350 ?     lit_ematn_dummy[] = lit_ematn[].

  • OPP(Output Post Processor) not processing the report (XML Publisher)

    Hi,
    I have defined a concurrent program (XML Publisher report) then ran but failed with the errors below. I am running the report in Oracle E-Business Suite 11.5.10.2, thru a concurrent manager other than Standard Manager. My guess is that OPP(Output Post Processor) is not processing the request outputs coming from a different manager/work shift since requests ran thru Standard Managers are all OK.
    In the OAM(Oracle Applications Manager)-> OPP, there is only 1 process allocated for both Actual and Target. If we increase the number of processes, will it work?
    /home/app/oracle/prodcomn/temp/pasta19356_0.tmp:
    /home/app/oracle/prodcomn/temp/pasta19356_1.tmp: No such file or directory
    Pasta: Error: Print failed. Command=lp -c -dkonica4 /home/app/oracle/prodcomn/temp/pasta19356_1.tmp
    Pasta: Error: Check printCommand/ntPrintCommand in pasta.cfg
    Pasta: Error: Preprocess or Print command failed!!!
    Anybody who has experienced similar issue?
    Thanks in advance.
    Rownald

    Hello,
    Just an additional tests info. We have 2 concurrent managers which I think is affecting the XML report output - Standard Manager(running 24 hours) and a Warehouse manager (9am-4:15pm).
    When I run the report before or after the Warehouse manager workshift(9am-4:15pm), the output is fine - which means PDF is generated and Pasta printing is OK. However, when report is run between 9am-4:15pm, the report only shows XML output with an error in Pasta printing(above). I also found that re-opening the output(ran prior to Warehouse workshift) on the period between 9am-4:15pm would also result to just an XML output instead of the previous PDF.
    Anybody who has experienced similar issue like this? Any idea? The report is not directly defined as "inclusion" in the Warehouse manager, only the program calling it. Any effect of multiple concurrent managers in the XML Publisher output?
    Thanks in advance for your ideas..

  • HT5631 how do I verify my apple id? I can't sign in to it in mail and I can't make a new one because it just will not process I'm trying to set up iCloud between iPad and iPhone and having ALOT of difficulty pleAse help need it done by later on today!!!!!

    how do I verify my apple id? I can't sign in to it in mail and I can't make a new one because it just will not process I'm trying to set up iCloud between iPad and iPhone and having ALOT of difficulty pleAse help need it done by later on today!!!!!

    In order to use your Apple ID to create an iCloud account, the primary email address associated with the ID must first be verified.  To do this, Apple will send a verification email to your account and you must respond to the email by clicking the Verify Now link.  Make sure you are check the spam/junk folder as well as the inbox.  If it isn't there, go to https://appleid.apple.com, click Manage your Apple ID, sign in, click on Name, ID and Email addresses on the left, then to the right click Resend under your Primary Email Address to resend the verification email.

  • Could not process due to error: com.sap.aii.adapter.file.ftp.FTPEx: 550

    Hi Experts,
    We have many File to EDI scenarios wherein XI System pick up the XML and sent to customers via EDI. Recently we faced a problem so created a Back-up System (Production copy) and tested successfully. After sometime the messages were routed to this back-up system and when we notice it and then stopped the back-up system. All the messages that were routed to back-up system, we try to send the same messages from the actually Production system to our customers. Now the problem is XI system (Production system) is unable to pick these files and I check the communication monitoring and encountered the below error message.
    Could not process due to error: com.sap.aii.adapter.file.ftp.FTPEx: 550.550
    Can anyone let me know how to fix the issue or what needs to be done?
    Your help is highly appreciated.
    Regards
    Faisal

    Hi,
    It seems to be problem with permission of files. Please ask your basis to do following:
    1. Set the permissions to FTP User you are using as 777 rights(full access to read , write and delete)
    2.If you have access to PI server, try to telnet /connect to ftp using command prompt (open ftp .....) the FTP server form there, you should see the same error there , inform this to your network guys.
    3.Clear all the files places already in the ftp (take backup) and test afresh after permissions are set by basis team.
    Regards
    Aashish Sinha

  • Regarding Rough FISCAL YEAR VARIANT NOT PROCESSED

    Hi Experts,
    I am loading data from 2lis_03_um to 0IC_C03,While executing DTP i am getting error as Rogh FISCAL YEAR VARIANT NOT PROCESSED.
    Can anyone please suggest me how to resolve the issue.
    Thanks
    Laxman

    Hi,
    Use the settings here:
    transaction code SPRO
    SAP Netweaver
    Business Intelligence
    Settings for BI content
    Trande Industries
    Retailing
    => set fiscal year variant
    regards,
    pascal

  • Bill and Routing Interface concurrent program is not processing component

    Hi All,
    While running Bill and Routing Interface concurrent program,not processing the component item(Line Item) from bom_inventory_comps_interface,still it show process flag as 1. where as billing item(Header Item) is successfully done..i have tried with all the option by giving component sequence id and batch id still it is not processing component item..
    Is it recommendable to give bill sequence id and component sequence id in interface table using bom_inventory_component_s sequence..
    I will be very pleased if i listen from u guys... Plz help me in resolving this issue..
    My Header Insert Stmt..
    INSERT INTO apps.bom_bill_of_mtls_interface@system_link_visma
    (assembly_item_id, organization_id,
    alternate_bom_designator, last_update_date,
    last_updated_by, creation_date, created_by,
    revision, item_number, item_description,
    implementation_date, transaction_type,
    process_flag, assembly_type, batch_id
    VALUES (l_inv_item_id, l_org_id,
    NULL, SYSDATE,
    1318, SYSDATE, 1318,
    l_revision, l_item_num, l_description,
    SYSDATE, 'CREATE',
    1, 1, 10003535
    Component Insert stmt
    INSERT INTO apps.bom_inventory_comps_interface@system_link_visma
    (operation_seq_num, component_item_id,
    item_num, basis_type, component_quantity,
    auto_request_material, effectivity_date,
    disable_date, planning_factor,
    component_yield_factor,
    enforce_int_requirements,
    include_in_cost_rollup, wip_supply_type,
    supply_subinventory, supply_locator_id,
    check_atp, optional,
    mutually_exclusive_options,
    low_quantity, high_quantity,
    so_basis, shipping_allowed,
    include_on_ship_docs, required_to_ship,
    required_for_revenue, component_remarks,
    transaction_type, process_flag,
    assembly_item_id, component_item_number,
    batch_id,component_sequence_id
    VALUES (l_operation_seq, l_comp_item_id,
    cur2.item_sequence, l_basis, cur2.quantity,
    l_auto_request_mtls, cur2.from_date,
    cur2.TO_DATE, cur2.planning_factor,
    cur2.yield_factor,
    l_enforce_int_requirements,
    l_include_in_cost_rollup, l_supply_type,
    l_supply_subinventory, NULL,
    l_check_atp, l_optional,
    l_mutually_exclusive_options,
    cur2.minimum_quantity, cur2.maximum_quantity,
    l_sale_order_basis, l_shippable_flag,
    l_include_on_ship_docs, l_required_to_ship,
    l_required_for_revenue, cur2.comments,
    'CREATE', 1,
    l_inv_item_id, l_comp_item_num,
    10003535,apps.bom_inventory_components_s.nextval@system_link_visma
    For Subcomponent Insert Stmt
    INSERT INTO apps.bom_sub_comps_interface@system_link_visma
    (substitute_component_id,
    substitute_item_quantity,
    assembly_item_id, component_item_id,
    operation_seq_num, organization_id,
    substitute_comp_number,
    component_item_number,
    assembly_item_number,
    transaction_type, process_flag,
    enforce_int_requirements,
    effectivity_date,component_sequence_id,batch_id
    VALUES (l_sub_comp_item_id,
    cur3.quantity,
    l_inv_item_id, l_comp_item_id,
    cur2.operation_sequence, l_org_id,
    l_sub_comp_item_num,
    l_comp_item_num,
    l_item_num,
    'CREATE', 1,
    l_enforce_int_requirements,
    SYSDATE,apps.bom_inventory_components_s.currval@system_link_visma,10003535
    Thanks
    Raman Sharma
    Edited by: 929841 on May 4, 2012 12:28 AM
    Edited by: 929841 on May 4, 2012 2:58 AM

    You need to populate the organization_id or organization_code in bom_inventory_comps_interface.
    Here is a minimal insert
    INSERT INTO bom.bom_inventory_comps_interface
                (operation_seq_num, last_update_date, last_updated_by,
                 creation_date, created_by, process_flag, transaction_type,
                 bom_item_type,
                 effectivity_date, organization_code, assembly_item_number,
                 item_num, component_item_number, component_quantity
         VALUES (1                                                   -- op_seq_num
                 ,SYSDATE, 1433
                 ,SYSDATE, 1433, 1                                  -- process_flag
                 ,'Create',
                 4 -- bom_item_type 1 Model; 2 Option class; 3 Planning; 4 Standard; 5 Product family
                 ,SYSDATE - 1, 'PUB', 'SSGPARENT1'          -- assembly_item_number
                 ,10                                                     --item_num
                 , 'SSGCOMP1'                           -- component_item_number
                 , 10                                          --qty
                )Sandeep Gandhi

  • Lync 2013 client not processing history spooler folder (conversation history)

    We have been using Lync 2013 fully patched up to latest & greatest, along with Exchange 2013 for about 6 months now & everything is working
    great. (I did a AD migration from our old parent company in the UK as well as migrated exchange from 2010 to 2013, all that without a hitch for the most part.)
    the issue that seems to be popping up intermittently is conversation history. I manged to work around some of them for the most part & getting
    them to process; with the exception of 1 user.
    If he logs into another PC his conversation history works fine.
    However on his PC it does not seem to want to process the history spooler folder in "C:\Users\User\AppData\Local\Micros­oft\Office\15.0\Lync\[email protected]­m",
    there are 737 files in there (all of his convo history)
    Now what i did & it seems to sort of work is,Signed out of Lync, I deleted the users Sip folder after backing up the history spooler folder, logged
    back into Lync which recreated the sip_user folder; logged back out of Lync , copied the History Spooler folder back to the SIP folder, & signed into Lync.
    This took 10 files out of the history spooler folder, & then appeared in the convo history; both on Lync client & Outlook; but did not bring
    all of them in. I did some testing and if you log out, then back in, it processes a few more. but not all.
    EWS & MAPI settings & all other Lync configuration information settings match a users with no convo history issues.
    So now the questions. 
    1)What could possibly cause what ever mechanism that processes the History spooler folder to stop/only process a few items during sign in? Also what
    is the mechanism? 
    2)Is there any way through powershell to force the synchronization of the history spooler folder?
    Thank you

    so in some more testing i had check the history spooler folder, & there were about 700+ files.
    as a note* (Upon deleting the users sip folder & logging back in thus re-creating the folder new convo's where saving. So i did "fix" the issue with no new convo's during my initial work prior to opening this thread.)
    I then removed a good chunk of the *.hist files & only left convo's from this month. signed out then back in, 10 files processed. then did it again, another 10. did it again, & it did 4 then stopped. I noticed there were some files over 1 mb. &
    remember coming across this post (Lync
    not processing files over 1mb) so i removed the 1mb files & then logged out then back in. processed 10 more. 
    i then added all the files to the history spooler file with the exception of the files over 1 mb, & logged in & out about 70 times to process the whole folder. Now all of the convos are in Lync & outlook as well as new convo's being saved. However
    i have a folder now with 130 files that are over 1mb that will not process at all..
    so now i am trying to find some validity on the statement in the link above 
    "We did open a case with Microsoft and they confirmed it is a bug in the Exchange EWS HTTP stack. It
    is supposedly under review by the Dev team, but my guess is that it won't be fix soon. The magic size is exactly 1MB or 1048576 bytes."
    If that is actually true it sucks & is some B.S as it is nearly 2 months later & this is obviously
    a major problem & not even a CU or PU has been released to address the issue.

  • Certificate issues Active Directory Certificate Services could not process request 3699 due to an error: The revocation function was unable to check revocation because the revocation server was offline. 0x80092013

    Hi,
    We have some problems with our Root CA. I can se a lot of failed requests. with the event id 22: in the logs. The description is: Active Directory Certificate Services could not process request 3686 due to an error: The revocation function was unable to
    check revocation because the revocation server was offline. 0x80092013 (-2146885613).  The request was for CN=xxxxx.ourdomain.com.  Additional information: Error Verifying Request Signature or Signing Certificate
    A couple of months ago we decomissioned one of our old 2003 DCs and it looks like this server might have had something to do with the CA structure but I am not sure whether this was in use or not since I could find the role but I wasn't able to see any existing
    configuration.
    Let's say that this server was previously responsible for the certificates and was the server that should have revoked the old certs, what can I do know to try and correct the problem?
    Thank you for your help
    //Cris

    hello,
    let me recap first:
    you see these errors on a ROOT CA. so it seems like the ROOT CA is also operating as an ISSUING CA. Some clients try to issue a new certificate from the ROOT CA and this fails with your error mentioned.
    do you say that you had a PREVIOUS CA which you decomissioned, and you now have a brand NEW CA, that was built as a clean install? When you decommissioned the PREVIOUS CA, that was your design decision to don't bother with the current certificates that it
    issued and which are still valid, right?
    The error says, that the REQUEST signature cannot be validated. REQUESTs are signed either by itself (self-signed) or if they are renewal requests, they would be signed with the previous certificate which the client tries to renew. The self-signed REQUESTs
    do not contain CRL paths at all.
    So this implies to me as these requests that are failing are renewal requests. Renewal requests would contain CRL paths of the previous certificates that are nearing their expiration.
    As there are many such REQUEST and failures, it probably means that the clients use AUTOENROLLMENT, which tries to renew their current, but shortly expiring, certificates during (by default) their last 6 weeks of lifetime.
    As you decommissioned your PREVIOUS CA, it does not issue CRL anymore and the current certificates cannot be checked for validity.
    Thus, if the renewal tries to renew them by using the NEW CA, your NEW CA cannot validate CRL of the PREVIOUS CA and will not issue new certificates.
    But it would not issue new certificates anyway even if it was able to verify the PREVIOUS CA's CRL, as it seems your NEW CA is completely brand new, without being restored from the PREVIOUS CA's database. Right?
    So simply don't bother :-) As long as it was your design to decommission the PREVIOUS CA without bothering with its already issued certificates.
    The current certificates which autoenrollment tries to renew cannot be checked for validity. They will also slowly expire over the next 6 weeks or so. After that, autoenrollment will ask your NEW CA to issue a brand new certificate without trying to renew.
    Just a clean self-signed REQUEST.
    That will succeed.
    You can also verify this by trying to issue a certificate on an affected machine manually from Certificates MMC.
    ondrej.

  • BPM process switch/fork is not processing!!

    Hi Guys,
    I have scenario for 1:n multi mapping using BPM. In my BPM just I have got 5 steps, the first one is receiver,  next step is transformation (workinig fine), switch with 2 branches, on branch one Sender 1 to SNC (Prod Active Notific) and on second branch has got Sender 2 for (ProductDemandInfluencingEventNotification) to SNC. The process is coming upto Switch, the process is not coming into Swithch step to send these messages to sender? what is the condition for Switch step? I put it like (1=1), message is not processing completely.
    Any valuable inputs for this please?
    Many Thank in advance
    San

    Hi biplab das ,
    Many thanks for very quick reply. I have tried both. Let me explain you my scenario, please give me your inputs on that!!
    Sender side xml messages comes with couple of messages. What I need this xml has to split into two message types "Product Activity Data" and "Product Influence Demand", each message goes with multiple messages, as follows:
    <Root>
       <output>
       <output>
       <output>
       <output>
       <output>
    </Root>
    Each output message got multiple submessags. Couple of submessages has go under "Product Activity Data" and Couple of other messages has go under "Product Influence Deman"
    Mapping is working fine. Messages are splitting.  I have configure BPM as follows:
    Setp 1---- Receiver --- > Receive the message
    Setp 2 -
    Mapping -
    > Messages are splitting
    Step 3 -
    Fork Step ---> with 3 Branches , Branch 1 for, 1 Receiver Determination and 1 Sender --> Product Acitivity Data
                                                                           Branch 2 for , 2nd Receiver Determination and 2nd Sender --> Product Influence D
                                                                           Branch 3 for throw Alert
    I can this message is failing on Moni, I opened PE, the graphical represenation shows error on Sender1 and Sender2.
    Please can you advice me in this.
    Many Thanks
    Regards
    San

  • IDOC is not processing in R/3 with status 64(Urgent)--Production issue.

    Dear Folks,
    Here we are implementing MIRROR sytem for data extraction in to BIW.
    I had gone through the steps in PDF file(How to minimize down thime for Delta Initialization) which has given by SAP.
    These are following steps I had performed in BW production
    1. Using RSADMIN transcation- Assigned one user as a Debugging user.
    2. With the Debugging user Created one infopackage on Purhcasing ITM datasource--
        In data selection tab, I given date as from 2002 to 2003
        In Update tab, I selected update mode as Initialize Delta process-               without data transfer and Uncheck Update Data in master   system immediately
       In schedule tab I selected option Start Data load immediately then Start extraction.
    3.Data was requested and It was intialized properly with green status 1 from 1 Records.
    4. In details tab Request, Extraction, transfer(IDOC and TRFC), Processing everything in Green color.
    5.As per document I have checked in R/3 for IDOC with status 64 using transaction BD87/WE05
    Problem is that IDOC is not processing in R/3 system. for that IDOC I have checked in BW montior screen using the option <b>IDOC list in Source sytem</b>, here also it is not showing any IDOCs.
    Please let me know why BW sytem is not sending any IDOCs to source sytem.Is there any settings I could have miss in BW aswell as in R/3 system.
    Note: In BW production I successfully loaded master data from R/3. aswell as I triggered infopackage-Initialization with data transfer, it is showing IDOC number properly in R/3 system.
    Points is Assured for all replies.

    Hi,
    - check you RFC destination (normal test + authorization test)
    - Check you partner profiles in WE20 and its ports WE21
    - from the source system tree (RSA13) right click your source system and perform a check
    - the best is to restore your source R/3 source system from BW (right click on it ans select RESTORE); you'll need to be able to log in the source system as an admin as well as opening the source system for customizing in SCC4 so that BW can recreate the correct partner parameters automatically; the best is to do that with your basis group.
    hope this helps,
    Olivier.

  • Script Alert: "Sorry I could not process the following files" (Error Message using Image Processor)

    I'm a seasoned Photoshop/Bridge CS5 user who recently upgraded to CS6.  In Bridge I just ran my first Image Processor batch, trying to convert a set of RAW .NEF files to .PSD files with a basic editing action I created applied to them.  Photoshop CS6 opens the files from Bridge and applies the actions, but then I get the "Script alert" error message "Sorry I could not process the following files" and a list of all of the files in the batch I was trying to run the script on.  The files all remain open and are saved on my hard drive in a folder created by the Image Processor script, but do not close unless I manually close them.  I never had this problem in CS5 where the image processor would apply the desired actions to the batch of photos one at a time, then save and close.  Thoughts on how I can fix this bug?

    Hey Jeff--
    No specific action steps to make the image safe - no flattening, etc. Had problems narrowing down the specific action step causing the problem. So I ran the Image Processor script in the ExtendScript ToolKit and debugged in. (I'm using the version that shipped with CS4.)
    Turns out what was happening is that the app.activeDocument.close() command on line 1490 was never getting called, because the this.SaveFile() above it was throwing an exception. More debugging got me to the problem - SaveFile() tries to restore the history state after making a doc safe for saving as a JPEG. This causes an exception. So I commented out line 1571:
    // app.activeDocument.ativeHistoryState = historyState;
    The script runs just fine after this. I commented out the close to see what history states got added to the doc, but no new states are recorded, so commenting out the restore line should not cause any problems.
    Of course, my solution does not solve the problem and I may need to comment out all the saveFile() lines that does this. Not sure why some actions have this problem and others don't. The actions I am running have about 25 steps. Perhaps PS is not catching up quick enough. I added a $.sleep(2500) after the restore history but that didn't help.
    So, not sure where this gets us. I will check about posting an action set with just one of the actions that cause the problem. But I may also have our customer comment out that line of the Image Processor script.
    Hope this helps. May post more next week.

  • HT4623 I can not get connected to App Store , download or update any apps,  EVEN I DID RESET ALL SETTING , STILL I RECIEVE A MESSAGE says: YOUR REQUEST IS NOT PROCESSED ,ERROR COD: 109 !! WHAT DOSE IT MEAN? HOW CAN I SOLVE THE PROBLEM??

    I can not get connected to AppStore , DOWNLOAD OR UPDATE ANY OF MY APPS , WHEN I GO FOR INSTALLING , THE MESSAGE SAYS: YOIR REQUEST IS NOT PROCESSED ,  ERROR CODE, 1009!!! HOW CAN I SOLVE THIS PROBLEM??

    Try contacting the iTunes support staff, they do the app store also, at:  http://www.apple.com/emea/support/itunes/contact.html

  • Email/print in status "Not processed" when RFQ is published

    Hello,
    We are running SRM 5.0.
    We have set up the event schema for BUS2200 (RFQ) to react on events: PUBLISHED, PUBLISHED_AGAIN etc. And assigned the event schema to RFQ's transaction types, eg BID.
    However, when an RFQ is published, the email is not sent to vendor (whose default medium is MAI) or printed on vendor's default printer (if the medium is set to PRN). The output log under bidder/bids tab shows something like :
    Status: Not processed       
    Medium: Print      
    When I clicked on the action button "Output pushbutton", the status would be changed to "Successful" and the email would be sent or the print job would be printed then.
    I have checked the PPTTRIGG table and all the entries in there are of DISPATCH = 1 or 4. Not 3, which seems to indicate an "immediate processing".
    My question is how I would configure the system so that the email/print when the RFQ is published could be processed immediately.
    Thanks.

    Hello Masa,
    I realized it as well. I was then a little puzzled why SRM didn't allow "immediate processing" for the "processing time" parameter in the action definition. Do you know if there is a particular reason for that?
    Thanks,

  • Queues are not processed properly in BPM

    Hi all ,
    In BPM , the inbound queues of the process engine are not processed properly.
    Several message are sent to the process and the message queued in the QRFC queue.
    How to process those messages so that process instance gets created ?
    Kind regards
    Goutam

    Check steps I've detailed on this thread:
    Re: problem in triggering BPM engine
    Hope that helps
    James.

Maybe you are looking for

  • Dynamic sql and block based on a stored procedure

    Hi! I'm triying to generate a block based on a stored procedure. I want the stored procedure to execute a dynamic sql, but I'm having getting the error "PLS-00455 cursor 'AULOCASLATE' cannot be used in dynamic SQL OPEN statement". I have the followin

  • How do I print a copy of my purchus history

    how do I print a copy of my purchus history

  • Strange ALV behaviour with dropdowns

    Hello, I've got an ALV grid control (cl_gui_alv_grid) that displays dropdowns (not F4 but a dropdown set via the set_drop_down_table method) in one column. It works fine in many cases but now I noticed the following: first I pass a table filled with

  • OBIO connection problem

    Hi, I used the oracle beehive integration for outlook downloader and installed OBIO. When I try to log into the server I get "Invalid port specified for the direct connection. Please try again". Can anyone tell me what I'm doing wrong? Which port is

  • IPod doesn't respond

    ive had my ipod for over a year and a half, and i know age can have an effect, but i was using my ipod for data storage about a week ago and i tried to turn it on, and it wouldnt, even though it was fully charged. i also tried to sync it with the com