CK24 Marking Batch Jobs causes SM12 Record locks

When finance runs the CK24 Marking job at month end, it seems to leaving some record locks hanging in SM12. The batch job is sucessful and says it has processed 23,699 materials, but it always leaves locks on some 700~ materials.
The materials are rohs, halbs, and ferts. I cannot figure out why only some cause this problem.
I am able to cleanup the locks, by deleting in SM12, but I want to know how to prevent, because it stops all functions such as shipping on these materials.
Thanks,
Bev

Dear,
Please check SM37 is for job logs
Have you "enqueue" in RZ20 ?
In SM12 By double-clicking a lock entry, you can display detailed information, including the host name and number of the SAP System in which the lock was generated.
check this sap help...
http://help.sap.com/saphelp_nw04/helpdata/en/7b/f9813712f7434be10000009b38f8cf/frameset.htm
Regards,
R.Brahmankar

Similar Messages

  • Can a long running batch job causing deadlock bring server performance down

    Hi
    I have a customer having a long running batch job (approx 6 hrs), recently we experienced performance issue where the job now taking >12 hrs. The database server is crawling. Looking at the alert.log showing some deadlock,
    The batch job are in fact many parallel child batch job that running at the same time, that would have explain the deadlock.
    Thus, i just wondering any possibility that due to deadlock, can cause the whole server to be crawling, even connect to the database using toad is also getting slow or doing ls -lrt..
    Thanks
    Rgds
    Ung

    Kok Aik wrote:
    According to documentation, complex deadlock can make the job appeared hang & affect throughput, but it didn't mentioned how it will make the whole server to slow down. My initial thought would be the rolling back and reconstruct of CR copy that would have use up the cpu.
    I think your ideas on rolling back, CR construction etc. are good guesses. If you have deadlocks, then you have multiple processes working in the same place in the database at the same time, so there may be other "near-deadlocks" that cause all sorts of interference problems.
    Obviously you could have processes queueing for the same resource for some time without getting into a deadlock.
    You can have a long running update hit a row which was changed by another user after the update started - which woudl cause the long-running update to rollback and start again (Tom Kyte refers to this as 'write consistency' if you want to search his website for a discussion on the topic).
    Once concurrent processes start sliding out of their correct sequences because of a few delays, it's possible for reports that used to run when nothing else was going on suddenly finding themselves running while updates are going on - and doing lots more reads (physical I/O) of the undo tablespace to take blocks a long way back into the past.
    And so on...
    Anyway, according to the customer, the problem seems to be related to the lgpr_size as the problem disappeared after they revert it back to its orignial default value,0. I couldn't figure out what the lgpr_size is - can you explain.
    Thanks
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    "Science is more than a body of knowledge; it is a way of thinking" Carl Sagan

  • Batch Jobs fail because User ID is either Locked or deleted from SAP System

    Business Users releases batch jobs under their user id.
    And when these User Ids are deleted or locked by system administrator, these batch jobs fail, due to either user being locked or deleted from the system.
    Is there any way that these batch jobs can be stopped from cancelling or any SAP standard report to check if there are any batch jobs running under specific user id.

    Ajay,
    What you can do is, if you want the jobs to be still running under the particular user's name (I know people crib about anything and everything), and not worry about the jobs failing when the user is locked out, you can still achieve this by creating a system (eg bkgrjobs) user and run the Steps in the jobs under that System User name. You can do this while defining the Step in SM37
    This way, the jobs will keep running under the Business User's name and will not fail if he/she is locked out. But make sure that the System User has the necessary authorizations or the job will fail. Sap_all should be fine, but it really again depends on your company.
    Kunal

  • Recorded Actions and Batch jobs

    Photoshop has an Actions panel feature that allows you to record and save a set of commands (eg: resize an image to 400 x 600, then turn it black & white).  Anytime you want to perform the same set of commands, just select and run the saved Action.  In addition, Photoshop has a Batch jobs feature that allows you to select a directory of files and run the saved action on all files in the directory.  In this way, you could resize a thousand images to 400 x 600 and turn them all black & white with one click.
    Does Captivate have any similar capability?  I have about 3,000 swf files and I wish to set the preloader percentage on all of them to the same value. I would hate to open, edit, save, and close the files manually one at a time.  If this feature does not exist, can you guys think of an efficient way to accomplish the task?
    I have Captivate 5 but can upgrade if necessary.
    Thanks for the help

    Since you mention that you have both CPT and CPTX files it might be worth mentioning that your goal of just opening them, chaning the preloader percentage before save and close again might not work out as planned. 
    Have you considered the following issues?
    Are you intending to open the CPT files in the same version of Captivate that created them, or were you thinking of upgrading them to a later version such as Cp5.5?  If so, you'll need to do more than just change the preloader percentage.  You'll also need to change the path to the preloader itself since each version of Captivate is installed in a different directory in Program Files. 
    Were all of the CPT projects set up as AS3 with all AS3 components, or were some of them AS2 and therefore would need to be reviewed and tested to make sure upgrading to a Cp5x version did not have other impacts?  Projects that were set up as AS2 would need to change preloaders, skins, and any animations over to AS3 before the upgrade would be successful.
    CPT to CPTX files will also experience font size issues since the font rendering technology changed in Cp5x.  So you'll probably need to go through those projects and fix a lot of issues in text captions.
    And another one to watch out for:  After upgrading projects from CPT to CPTX the Continue button on the Quiz Results will often cease to work unless you remove and replace the Quiz Results slide.
    This is not a complete list of potential upgrade issues.  What I'm trying to sound here is a warning that, unless you use exactly the same version of Captivate that created them, making any changes to these files could turn into a much bigger job than you hope.  There could be weeks of work involved trawling through each project file one by one checking for issues.

  • Output condition records copied by batch job

    Hi!
    We are creating invoices through batch job RV60SBAT. However, we notice that the outputs are created based on the output condition records of the last invoice created by the batch job.
    For example, if 2 invoices were to be picked up for creation by the batch job, where the first invoice should create a print output and the second invoice should create an email output, both invoices will be created with email outputs.
    It appears that the codition record of the last invoice is copied to all invoices created by a batch job run.
    Has anyone experienced the same issue?
    Cholen

    first invoice should create a print output and the second invoice should create an email output
    In that case, create two variants in program SD70AV3A by differentiating the output type and schedule it in background
    ps:-  post your text in big fonts as it is difficult to read
    G. Lakshmipathi

  • Table lock in Batch jobs

    There is a program which runs in batch mode every 30 mins. After successful execution, program updates last run time stamp in a Z table. For each sales org we set one batch job, like this there are several jobs running in same time & some jobs are not updating time stamp. My gut feeling is reason could be table lock. We canu2019t ask batch team to schedule  all jobs in sequential manner with in 30 mins duration because of huge number of jobs.
    How we can overcome this problem.. Check for table lock using ENQUEUE_E_TABLES and wait for some time is feasible solution? Can some one guide me in this regard please?
    Thanks in advance.

    Hi Perez C,
    you may try ENQUEUE and skip for reprocessing, you can also set WAIT = 'X' in the  ENQUEUE call. This will wait a pre-set time (default I think like 3 seconds) trying to lock the object. In most cases it helps. But if you can not lock successfully, you should write it to application log at least.
    Regards,
    Clemens

  • Duplicate deliveries getting created by batch job against STO

    Hi Experts,
    I am facing one issue where duplicate deliveries are getting created by batch job against Intercopmany STO.
    Scenario is PO having one line item with ordered qty of 8000kg.
    Through batch job, two deliveries got created on same day for PO line item.
    One delivery got created for 8000kg and another delivery has got created for 7000Kg. So user has deleted second delivery of 7000kg.
    Next day again the delivery got created for 8000kg for the same PO line item through batch job.
    I am wondering how the duplicate deliveries are getting created by batch job for PO even though it has no open items.
    All deliveries got created through batch job only as cross checked the user name in delivery.
    Kindly help to fix the issue.

    Hi Amit
    I assume you are talking about outbound deliveries.  In this case it would be worth checking the customer master record for the receiving plant.  In the sales area data there is a shipping tab which contains several settings used to control delivery creation for customers.
    It is possible to control how the system behaves when you have a stock shortage and restrict the number of partial deliveries.  This might help you control this situation and might be the cause.
    Regards
    Robyn

  • How to find out batch job failure and taking action:

    Normally We will monitor the batch jobs  through transaction code sm37 for job monitoring. In SM37 we will give a batch job name date and time  as  input. In the first step we will check the batch job for the reason failure or check the spool request for the batch job for failures an help in analyzing the error
    I understand from the my experience is that the batch may fail due to below reasons.
    1.,Data issues :             ex: Invalid character in quantity (Meins) field  >>>> We will correct the corresponding document with correct value or we will manually run or request the team to rerun the batch job by excluding  the problematic documents from the batch job variant  so that it may process other documents.
    2.Configuration issues : Materials XXXX is not extended for Plant >>>> we will contact the material master team or business to correct the data or we will raise sub contract call with support team to correct he data. Once the data been corrected and will request the team to rerun the batch job.
    3.Performance issues : Volume of the data being processed  by the batch job ,network problems.>>>Normally these kind of issues we will encounter during the month end process as there will lot of accounting transactions or documents being posted business hence it may cause the batch job failure as there is enough memory to complete the program or select queries in the program will timeout because of volume of the records.
    4.Network issues. : Temporary connectivity issues in other partner systems :Outage in other partner systems like APO or other system like GTS  will cause the batch job failure as Batch job not in position to connect other system to get the inforamtion and proceed for further steps.Nornmally we will check RFC destination status by running a custom program  weather connectivity between system are in progress or not. then intimate other partner system  for the further actions, Once the partner system comes online then we will intimate the team to restart or manually submit batch job.
    Some times we will create a manual job by transaction code SM36.

    I'm not sure what the question is among all that but if you want to check on jobs that are viewable via SM37 and started via SM36. The tables are TBTCP -Background Job Step Overview and TBTCO - Job Status Overview Table.
    You can use the following FM to get job details:
    GET_JOB_RUNTIME_INFO - Reading Background Job Runtime Data

  • FOR UPDATE cursor is causing Blocking/ Dead Locking issues

    Hi,
    I am facing one of the complex issues regarding blocking / dead locking issues. Please find below the details and help / suggest me the best approach to ahead with that.
    Its core Investment Banking Domain, in Our Day to day Business we are using many transaction table for processing trades and placing the order. In specific there are two main transaction table
    1)     Transaction table 1
    2)     Transaction table 2
    These both the tables are having huge amount of data. In one of our application to maintain data integrity (During this process we do not want other users to change these rows), we have placed SELECT …………….. FOR UPDATE CURSOR on these two table and we have locked all the rows during the process. And we have batch jobs (shell scripts ) , calling this procedure , we will be running 9 times per day 1 hrs each start at 7:15AM in the morn finish it up in the eve 5PM . Let’s say. The reason we run the same procedure multiple times is, our business wants to know the voucher before its finalized. Because there is a possibility that order can be placed and will be updated/cancelled several times in a single day. So at the end of the day , we will be sending the finalized update to our client.
    20 07 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 08 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 09 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 10 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 11 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 12 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 13 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 14 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 15 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 16 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    20 17 * * 1-5 home/bin/app_process_prc.sh >> home/bin/app1/process.out
    Current Program will look like:
    App_Prc_1
    BEGIN
    /***** taking the order details (source) and will be populate into the table ****/
    CURSOR Cursor_Upload IS
    SELECT col1, col2 … FROM Transaction table1 t 1, Source table 1 s
    WHERE t1.id_no = t2.id_no
    AND t1.id_flag = ‘N’
    FOR UPDATE OF t1.id_flag;
    /************* used for inserting the another entry , if theres any updates happened on the source table , for the records inserted using 1st cursor. **************/
    CURSOR cursor_update IS
    SELECT col1, col2 … FROM transaction table2 t2 , transaction table t1
    WHERE t1.id_no = t2.id_no
    AND t1.id_flag = ‘Y’
    AND t1.DML_ACTION = ‘U’,’D’ -- will retrieve the records which are updated and deleted recently for the inserted records in transaction table 1 for that particular INSERT..
    FOR UPDATE OF t1.id_no,t1.id_flag;
    BLOCK 1
    BEGIN
    FOR v_upload IN Cursor_Upload;
    LOOP
    INSERT INTO transaction table2 ( id_no , dml_action , …. ) VALUES (v_upload.id_no , ‘I’ , … ) RETURNING v_upload.id_no INTO v_no -- I specify for INSERT
    /********* Updating the Flag in the source table after the population ( N into Y ) N  order is not placed yet , Y  order is processed first time )
    UPDATE transaction table1
    SET id_FLAG = ‘Y’
    WHERE id_no = v_no;
    END LOOP;
    EXCEPTION WHEN OTHER THEN
    DBMS_OUTPUT.PUT_LINE( );
    END ;
    BLOCK 2
    BEGIN -- block 2 starts
    FOR v_update IN Cursor_Update;
    LOOP;
    INSERT INTO transaction table2 ( id_no ,id_prev_no, dml_action , …. ) VALUES (v_id_seq_no, v_upload.id_no ,, … ) RETURNING v_upload.id_no INTO v_no
    UPDATE transaction table1
    SET id_FLAG = ‘Y’
    WHERE id_no = v_no;
    END LOOP;
    EXCEPTION WHEN OTHER THEN
    DBMS_OUTPUT.PUT_LINE( );
    END; -- block2 end
    END app_proc; -- Main block end
    Sample output in Transaction table1 :
    Id_no | Tax_amt | re_emburse_amt | Activ_DT | Id_Flag | DML_ACTION
    01 1,835 4300 12/JUN/2009 N I ( these DML Action will be triggered when ever if theres in any DML operation occurs in this table )
    02 1,675 3300 12/JUN/2009 Y U
    03 4475 6500 12/JUN/2009 N D
    Sample output in Transaction table2 :
    Id_no | Prev_id_no Tax_amt | re_emburse_amt | Activ_DT
    001 01 1,835 4300 12/JUN/2009 11:34 AM ( 2nd cursor will populate this value , bcoz there s an update happened for the below records , this is 2nd voucher
    01 0 1,235 6300 12/JUN/2009 09:15 AM ( 1st cursor will populate this record when job run first time )
    02 0 1,675 3300 12/JUN/2009 8:15AM
    003 03 4475 6500 12/JUN/2009 11:30 AM
    03 0 1,235 4300 12/JUN/2009 10:30 AM
    Now the issues is :
    When these Process runs, our other application jobs failing, because it also uses these main 2 tranaction table. So dead lock is detecting in these applications.
    Solutin Needed :
    Can anyone suggest me , like how can rectify this blocking /Locking / Dead lock issues. I wants my other application also will use this tables during these process.
    Regards,
    Maran

    hmmm.... this leads to a warning:
    SQL> ALTER SESSION SET PLSQL_WARNINGS='ENABLE:ALL';
    Session altered.
    CREATE OR REPLACE PROCEDURE MYPROCEDURE
    AS
       MYCOL VARCHAR(10);
    BEGIN
       SELECT col2
       INTO MYCOL
       FROM MYTABLE
       WHERE col1 = 'ORACLE';
    EXCEPTION
       WHEN PIERRE THEN
          NULL;
    END;
    SP2-0804: Procedure created with compilation warnings
    SQL> show errors
    Errors for PROCEDURE MYPROCEDURE:
    LINE/COL                                                                          ERROR
         12/9        PLW-06009: procedure “MYPROCEDURE” PIERRE handler does not end in RAISE or RAISE_APPLICATION_ERROR
         :)

  • Sessions from apex are blocking by batch job

    Hi All,
    In my 10.2.0.3 linux with apex 3.1.1
    We have a batch job using a package,that is doing DML operation in tables depends upon some condition.And this job will run daily 5 hours.
    The sessions from apex that is also using some procedures in the same package.Some days the sessions from apex are blocking by batch job.While blocking wait event from apex sessions are 'library cache pin'.Once the batch job completes then automatically removes the block.
    What may be the reason?
    Thanks in Advance,
    Sunil

    The query against v$lock is not relevant if the blocked sessions are waiting on "library cache pin" - so you seem to have two separate problems.
    If you have a session waiting on another session's transaction slot in mode 4 there are several possible causes - often related to indexes, but there are a couple of "internal" problems as well. If you see this locking issue again some of the simplest things to check for would be session 1 inserting (without commit) some rows in a table with a unique key, and session 2 then trying to insert a duplicate. Session 2 has to wait for session 1 to commit or rollback before deciding whether to return a "duplicate key" error, or to continue processing.
    The "library cache pin" waits suggest that the package had become invalid while the batch job is running it, and one of the Apex jobs is trying to recompile it. But if the batch job is currently running (hence pinning) the "executable", then the Apex job can't get the necessary exlusive pin until the batch job ends and releases its pin.
    Read the notes in the script $ORACLE_HOME/rdbms/admin/catblock.sql about creating views that let you see more of the information about library cache (KGL) locks and pins.
    Regards
    Jonathan Lewis
    http://jonathanlewis.wordpress.com
    http://www.jlcomp.demon.co.uk
    To post code, statspack/AWR report, execution plans or trace files, start and end the section with the tag {noformat}{noformat} (lowercase, curly brackets, no spaces) so that the text appears in fixed format.
    "Science is more than a body of knowledge; it is a way of thinking"
    Carl Sagan                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Package processing in CCM (AUTH_UPDATE batch-job)

    Hello all
    We're using CCM 2.0 with SP03. We send CatalogueUpdateNotification_in messages (in XML format) to our XI and route it 1:1 to our CCM system. There it is processed by an <b>async started AUTH_UPDATE batch-job</b>. This job finally processes our catalog items.
    As log as this batch-job runs the concerned catalog being updated is locked for other updates. And this is what causes us some head aches!
    We send XML messages containing up to 2000 catalog items. This makes the batch job run quite long (up to 1.5h). Now, when we have more than 2000 updates to send to our catalog, we split it up into packages with up to 2000 items each.
    Each package being sent from XI to CCM creates such a AUTH_UPDATE batch-job. The 1st starting is able to lock the catalog but this way of course also blocks processing of all other packages.
    Every single minute the next (waiting / blocked) package tries to lock the catalog itself before it can be processed. This ensures parallel update processings but the other way around it also causes to fall into a timeout after lets say ca. 30 retries.
    From XI point of view the timedout out package(s) sent to CCM were successfully processed... means: not restartable!
    In CCM I can't find a way trying to re-process such a timedout package.
    Re-generating the same package from the source system is not an option for us as well... it takes to much efforts.
    Is there a way to change the wait time between retrying to lock the catalog?
    OR
    Is there a way to ensure that only 1 AUTH_UPDATE batch-job is started at the same time?
    Thanks for some ideas to solve this.
    Renaud

    Hi
    <b>Please go through the SAP OSS Notes and pointers as well -></b>
    Note 908054 - CCM 2.0 SP03 on Netweaver 04s SP Stack 06
    Note 1002775 - Asynchronous deletion of old version after publication
    Note 955096 - Rule condition (Not Equal with Not Valuated) in the view
    Note 887328 - CCM20 - modifications for /CCM/VERIFY_CATALOG
    Note 865138 - CCM20SP0 upload overview message "Category not found"
    Note 873340 - Problems with BMECat mapping after patching XI
    Note 864313 - Input help for mass editor
    <u>Related links -></u>
    Re: Transfer Contract Data to CCM
    Re: CCM CSE_ENRICHMENT BAdI debuging
    Incase this does not help, please raise a customer OSS Message with SAP as well for faster resolution of your queries.
    Do let me know. Hope this will help.
    Regards
    - Atul

  • Logging Record Locks in 9i

    hi everyone,
    We have a 9i database that sometimes experiences several times per day where the user tries to save a work order only to have to app crash with a 'record in use' error.
    I searched \bdump\alert to see if I could find any loging for locks - didn't find any.
    Are severe record locks - severe enough to cause an update to fail - logged to a file? looking for a historical "hard-copy" because when I've looked at dba_blockers, v$locked_object etc I don't see any at that time.
    Thanks,
    John

    From an Oracle perspective, there is no such thing as a row-level lock that would cause an application to crash. I'm also not aware of any Oracle error with the words "record in use" in it, so this is probably an application-specific error and an application-specific crash. If this is a custom app, you probably want to modify the code so that it doesn't crash when a row is locked.
    Logging every lock of a row would be a substantial burden on scalability, so I don't believe there is any option to do this. If the application waits long enough to acquire the lock, you may be able to set up a job to query DBA_BLOCKERS/ DBA_WAITERS frequently enough to catch the error. But if the application isn't waiting for the lock, that's probably not going to work.
    Justin

  • Two Deliveries did not go through Batch Job Runing

    Hi,
    In this company every evening the automatic batch Job for Billing due list is run but on Nov 23 one Deliver and on Nov 30 another delivery has not gone through i.e. the invoices were not created for these two deliveries.
    I checked in VF01, the individual invoice is being created. The user is asking the reason for this and how to prevent this type of situation in future.
    I don't know the Batch Job number and I think it has been done in VF06.
    Please guide me how I can solve this issue.
    Thanks,
    Jans

    Hello Tushar,
    Thank you for your reply. I don't know the Variant name also and I don't know the Batch Job number eirther. The user is not aware of these things. I tried to find out from VF06 but I could not figure out anything.
    Please explain me your sentece "Execute billing program in foreground and check". you mean to say after creating the same scenario in the quality, I will run the Billing program ( I will take the Billing program from the VF01). If I am correct please let me know.
    Regarding debugging,  if I have to do how I can do it. I know to go to Debuggiing mode if I put in the command field  "/h" , I will be in the debugging mode. But my question is which transaction Should I use VL02N or VF01 to find out the Problem
    I have mentioned earlier that I went to VF01 and these deliveries are fine to create invoices but the User is interested to know the Problem and to find out the cause so that they can prevent it occuring in future.
    Thanks,
    Jans

  • Problem in creating Batch job for programs in different servers.

    Hi All,
    I have a requirement like i have 2 different programs ( PRG1,PRG2 ) running in 3 different SAP servers (SER1,SER2,SER3).
    I want to create a batch job for the programs.
    Here When I run the PRG1 in Server SER1,it will create a file1 in SER1 as well as the other 2 servers (SER2 and SER3).
    It just create a file and there is no content in the file.Once i start writing records to file1in SER1,the same record will be started writing in SER2 and SER3.This is using PRG1.
    Once PRG1 will complete writing records in file,Next PRG2 will run in SER2 using file created by PRG1.
    I want to create a batch job for PRG1.
    How to trigger the PRG2 after completing job1 for PRG1 without any manual interaction.
    Please help me for this.

    Hi Ambili
    You can set steps in these jobs in transaction SM36.
    You can define that Job 2 starts after Job1 complete sucessfully and same game for other jobs.
    Please, check attachment for detailed information about that
    Kind regards
    Edited by: Carlos Machado on May 27, 2010 6:01 PM

  • Launch batch job thru SOAP call : no execution, connection OK.

    Hello,
    I am experiencing some problems launching batch jobs thru Web Services. I have added batch job "EPN_Test_Webservices" to the Web Services and enabled the Job Attributes.
    When I try to launch the job thru a SOAP call, I get a reply but the job is not executed.
    This is the SOAP envelope:
    <SOAP-ENV:Envelope xmlns:SOAP-ENV="http://schemas.xmlsoap.org/soap/envelope/" xmlns:SOAP-ENC="http://schemas.xmlsoap.org/soap/encoding/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xsd="http://www.w3.org/2001/XMLSchema">
         <SOAP-ENV:Body>
              <m:EPN_Test_Webservices_Job xmlns:m="http://www.businessobjects.com/DataIntegrator/ServerX.xsd">
                   <job_parameters>
                        <job_system_profile>String</job_system_profile>
                        <sampling_rate>10</sampling_rate>
                        <auditing>true</auditing>
                        <recovery>true</recovery>
                        <job_server>JS_ict</job_server>
                        <trace>String</trace>
                   </job_parameters>
              </m:EPN_Test_Webservices_Job>
         </SOAP-ENV:Body>
    </SOAP-ENV:Envelope>
    This is the reply I get:
    <soapenv:Envelope xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
         <soapenv:Body>
              <BatchJobResponse>
                   <pid>3888</pid>
                   <cid>13</cid>
              </BatchJobResponse>
         </soapenv:Body>
    </soapenv:Envelope>
    When I use the "ping" operation I get the following reply:
    <soapenv:Envelope xmlns:soapenv="http://schemas.xmlsoap.org/soap/envelope/" xmlns:xsd="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
         <soapenv:Body>
              <pingVersion>
                   <version>Business Objects Data Integrator Version 11.7.2.0</version>
              </pingVersion>
         </soapenv:Body>
    </soapenv:Envelope>
    ...which indicates the connection works. Also the "processed counter" increases, but the job does not get executed. There is nothing in the  job log, there is no result (the job is supposed to write a record in a database table). What can be wrong?

    If I run the following (on the machine I'm trying to set up the share)
    smbclient -L localhost -U%
    I got the following output
    Connection to localhost failed (Error NT_STATUS_CONNECTION_REFUSED)
    so I thought it might be something incorrect with the iptables side of things, however I haven't really touched that at all and it seems to look correct
    iptables -nvL
    Chain INPUT (policy ACCEPT 667 packets, 79977 bytes)
    pkts bytes target prot opt in out source destination
    Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
    pkts bytes target prot opt in out source destination
    Chain OUTPUT (policy ACCEPT 157 packets, 20724 bytes)
    pkts bytes target prot opt in out source destination
    So from my (very little) knowledge this appears correct (I think)... However it appears that something is blocking access somewhere.

Maybe you are looking for

  • How can I make an update of application with a different Apple ID?

    Hello, I have set up many iPad in my company with my Apple ID, however we have to do many update of application and iPad ask my Apple ID intead new Apple ID set it up on the Setting menu "iTunes". Could you pls help me? Thanks Vanderlei Gomes <Person

  • How to call excel macros programmatically in C#?

    Hi, I have a requirement where i need to call excel (2003) macros in C# program. Can anyone help me with a code snippet to do the same? The excel macro function takes two input parameters? how can the parameters be passed? Any code snippet to do the

  • Thousands of radial shadings

    I am writing a plotting package, and I would like to be able to display thousands of spheres on a page. I've successfully written code to use the radial shadings with an exponential function to make really nice looking spheres. When I plot a little o

  • Slow only in a few places

    I have a new FIOS connection upgraded from DSL. I have: 1 laptop with XP and a wireless G card, 1 laptop with Vista and a wireless G card, and I have 1 wired desktop. Anyway my problem is that on any of the systems the speed on facebook apps is slowe

  • [C] Get information about touchpad

    Hi, I'm completly new to writing Linux applications in C. In my program I want to enable/disable my touchpad based on whether or not a mouse is present. But I can't seem to figure out which library I need to use to get information about my touchpad a