Query in Background scheduling job...

Hi friends,
         Is it posiible to schedule a report which has additional screen developed by me(i.e SCREEN 1000 which is standard + SCREEN 9000 which is developed by me to  get the input from user) in background mode.
        When i tried to do it,I was not able to see any SPOOL REQUEST NO. in SM37 Transaction,subsequently I was not able to see the output!!!!!!!!
With Regards,
Kumaresh.

Hi,
Yes, u can schedule in b/g .wat ever may be the screens.
As u said no spool is creatde....?
have u execeuted the reoprt online without any b/g? is it displaying the output?
Debug the code wheteher the required selection parameters/data  is available in screen 9000
Revert back , so that i may help..
Regards,
Naveen

Similar Messages

  • Debugging of background scheduled job

    Hi guys,
    Can any one explain me 1 scenario why there is a need to debug background scheduled job..., rather then doing in background we can run the same job/report online and do normal debugging.
    thx in advance,
    Cheers,
    Harish

    Hi ARPAN,
    as long as the job is running, you can enter debuuger from SM50 process overview (program->debug) or SM37 Job Overview (Job->catch active Job).
    And to the initial question: Some messages are configured to be created when saving document (e.g. sales invoice). This will trigger a background process. If something goes wrong with smartform or sapscript form, it will cause an update error. This you may need to debug in background.
    Regards,
    Clemens

  • Change variant for background scheduled jobs

    Hi experts,
    I want to schedule the report RKO7KO8G in background.
    Two of the selection parameters of this report are Settlement period & Settlement year.
    This report should be executed in the background every month. For this the period parameter should change every month
    , like for example when the report executes in background in May then period should be 05, June - period should be 06
    and so on.
    i.e. the variant should change.
    For this how do i create a dynamic variant. I tried to create dynamic variant but it says that the parameter
    should be a date field. However, the settlement period field over has input from 001 to 999 & not date field.
    I want to schedule this report in background on a monthly basis and change its input parameter Settlement period
    as per the month.
    Please suggest.
    Regards,
    Priti.

    Links to sap.technical are NOT permitted. Even via Tiny.
    Hi,
    Have a look at the below tutorial;
    [Dynamic Variant|http://joyjitghosh.blogspot.com/2008_09_01_archive.html]
    Regards,
    Karthik D
    Edited by: Matt on May 27, 2010 10:57 AM - changed url to the source

  • Gui_download Function module is not working in Background Scheduling

    Hi Experts,
    Function modules like GUI_DOWNLOAD are not working in background scheduled jobs, It is raising an exception. Though it is working fine in foreground execution.
    Please provide me solution for this challenging task.
    you could be rated well for this question.
    Thanks,
    Rama Krishna.

    Hello,
    work with open dataset and close dataset...
    just loop over itab and transfer itab field for field into a string field separated by ;
    Here is an extract of coding:
    OPEN DATASET p_dath FOR OUTPUT IN TEXT MODE ENCODING DEFAULT.
      LOOP AT itab_pos.
        CLEAR satz.
        MOVE itab_pos-vbeln TO satz-vbeln.
        MOVE ';' TO satz-fill1.
        MOVE itab_pos-kunnr TO satz-kunnr.
        MOVE ';' TO satz-fill2.
        MOVE itab_pos-ktgrd TO satz-ktgrd.
        MOVE ';' TO satz-fill3.
        MOVE itab_pos-fkdat TO satz-fkdat.
        MOVE ';' TO satz-fill4.
        MOVE itab_pos-budat TO satz-budat.
        MOVE ';' TO satz-fill5.
        MOVE itab_pos-cpudt TO satz-cpudt.
        MOVE ';' TO satz-fill6.
        MOVE itab_pos-fkimg TO satz-fkimg.
        MOVE ';' TO satz-fill7.
        MOVE itab_pos-netwr TO satz-netwr.
        MOVE ';' TO satz-fill8.
        MOVE itab_pos-vrkme TO satz-vrkme.
        MOVE ';' TO satz-fill9.
        MOVE itab_pos-ktgrm TO satz-ktgrm.
        MOVE ';' TO satz-fill10.
        MOVE itab_pos-matnr TO satz-matnr.
        MOVE ';' TO satz-fill11.
        MOVE itab_pos-mattxt TO satz-mattxt.
        MOVE ';' TO satz-fill12.
        MOVE itab_pos-gewei TO satz-gewei.
        MOVE ';' TO satz-fill13.
        MOVE itab_pos-brgew TO satz-brgew.
        TRANSFER satz TO p_dath.
      ENDLOOP.
      CLOSE DATASET p_dath.
    It is not so comfortable as it is with GUI_DOWNLOAD but it works in background!

  • Job getting cancelled in background scheduling

    Hi,
        I have written a report  in the SAP HR system and it is to be scheduled in background. This program transfers the PERNRs from HR system to CRM system as partners in BUT000.The report works perfectly fine if it is run in foreground. When I run it in background the job gets cancelled and I get the job log having a message that it could not ascertain code page and the system exception ERROR_MESSAGE is raised. If anybody has any clues regarding this problem please help me.
    Thanks & Regards,
    Rahul.

    Hi,
    I Guess you are using container controllers in the Report, and all container contols requires help from Sap Gui for execution. I guess that is the reason it is failing in background . Please check on that aspect as well.
    Hope this helps..

  • Background Processing? how schedule job for "System Error" Message .

    Hello everyone,
    in sap help i have read.
    http://help.sap.com/saphelp_nw04/helpdata/en/5a/f72040599a8f5ce10000000a155106/frameset.htm
    PCK> Monitoring>Message Monitoring-->Background Processing
    you can schedule jobs for various background processing:
    ●     Archiving of messages processed successfully
    ●     Deletion of messages that are not to be archived
    ●     Restarting of messages with errors
    ●     Rescheduling of lost messages
    can anyone understand this docu?
    give me some introduction, how can i define and schedule these jobs ?
    thx in advance!!
    best regards
    Yaning

    Background Processing
    Prerequisites
    You have started the message monitor on the initial screen of the PCK and are in Background Processing.
    Features
    Archiving
    You require two archiving sessions to archive messages:
    ●     One session to write the messages to the archive
    ●     One session to delete the persisted messages that have been archived
    To do this, you schedule an archiving job, which implicitly schedules the sessions to write to the archive and delete the archived messages.
    You can define one or more rules for each archiving job; these rules contain conditions that a message must meet in order to be archived by the job. At least one of the defined rules must be met for archiving to take place.
    All information that is displayed for a message in message monitoring is archived, in addition to the audit log for each message.
    Deleting
    A standard delete job is created automatically. It runs once a day. You can schedule additional delete jobs; however, you cannot define rules for them.
    Restarting
    Instead of restarting messages with errors manually with message monitoring, you can schedule a job to automatically restart these messages. This is possible for all messages for which the number of defined restart attempts has been exceeded (messages with the system error status).
    You can define one or more rules for each job to restart messages; these rules contain conditions that a message must meet in order to be restarted by the job. At least one of the defined rules must be met for archiving to take place.
    Rescheduling
    A standard job to reschedule messages is created automatically. The job runs once a day and ensures that messages lost as a result of database failure, for example, are rescheduled. You can schedule additional rescheduling jobs; however, you cannot define rules for them.
    Thx Aamir.
    But I mean the messages with errors in Adapter Engine , not in Intergrations Engine.
    the situation is like Naveen Pandrangi's WebLog
    II. Errors in Adapter Engine [XI :  How to Re-Process failed XI Messages Automatically|XI :  How to Re-Process failed XI Messages Automatically]
    I
    Till now we have seen how to resubmit/restart message that failed in Integration Engine.  One a message makes it from Integration Engine to Adapter Engine, the message is flagged as checked in Integration Engine. The status of the message in Adapter engine does not effect the processed state in Integration Engine. Now if this message was asynchronous, XI will by default try to restart the message 3 times at intervals of 5 minutes before the status of the message is changed from Waiting to System Error .
    *how can i schedule a job to automatically restart these messages with errors?
    best regards
    Yaning
    Edited by: Yaning Liu on Aug 18, 2008 1:43 PM

  • Parallel processing in background using Job scheduling...

    (Note: Please understand my question completely before redirecting me to parallel processing links in sdn. I hve gone through most of them.)
    Hi ABAP Gurus,
    I have read a bit till now about parallel processing. But I have a doubt.
    I am working on data transfer of around 5 million accounting records from lagacy to R/3 using Batch input recording.
    Now if these all records reside in one flat file and if I then process that flat file in my batch input program, I guess it will take days to do it. So my boss suggested
    to use parallel processing in SAP.
    Now, from the SDN threads, it seems that we have to create a Remote enabled function module for it and stuf....
    But I have a different idea. I thought to dividE these 5 million records in 10 flat files instead of just one and then to run the Custom BDC program with 10 instances which will process 10 flat files in background using Job scheduling.
    Can this be also called parallel processing ?
    Please let me know if this sounds wise to you guys...
    Regards,
    Tushar.

    Thanks for your reply...
    So what do you suggest how can I use Parallel procesisng for transferring 5 million records which is present in one flat file using custom BDC.?
    I am posting my custom BDC code for million record transfer as follows (This code is for creation of material master using BDC.)
    report ZMMI_MATERIAL_MASTER_TEST
          no standard page heading line-size 255.
    include bdcrecx1.
    parameters: dataset(132) lower case default
                                 '/tmp/testmatfile.txt'.
       DO NOT CHANGE - the generated data section - DO NOT CHANGE    ***
      If it is nessesary to change the data section use the rules:
      1.) Each definition of a field exists of two lines
      2.) The first line shows exactly the comment
          '* data element: ' followed with the data element
          which describes the field.
          If you don't have a data element use the
          comment without a data element name
      3.) The second line shows the fieldname of the
          structure, the fieldname must consist of
          a fieldname and optional the character '_' and
          three numbers and the field length in brackets
      4.) Each field must be type C.
    Generated data section with specific formatting - DO NOT CHANGE  ***
    data: begin of record,
    data element: MATNR
           MATNR_001(018),
    data element: MBRSH
           MBRSH_002(001),
    data element: MTART
           MTART_003(004),
    data element: XFELD
           KZSEL_01_004(001),
    data element: MAKTX
           MAKTX_005(040),
    data element: MEINS
           MEINS_006(003),
    data element: MATKL
           MATKL_007(009),
    data element: BISMT
           BISMT_008(018),
    data element: EXTWG
           EXTWG_009(018),
    data element: SPART
           SPART_010(002),
    data element: PRODH_D
           PRDHA_011(018),
    data element: MTPOS_MARA
           MTPOS_MARA_012(004),
         end of record.
    data: lw_record(200).
    End generated data section ***
    data: begin of t_data occurs 0,
          matnr(18),
          mbrsh(1),
          mtart(4),
          maktx(40),
          meins(3),
          matkl(9),
          bismt(18),
          extwg(18),
          spart(2),
          prdha(18),
          MTPOS_MARA(4),
        end of t_data.
    start-of-selection.
    perform open_dataset using dataset.
    perform open_group.
    do.
    *read dataset dataset into record.
    read dataset dataset into lw_record.
    if sy-subrc eq 0.
    clear t_data.
    split lw_record
       at ','
    into t_data-matnr
          t_data-mbrsh
          t_data-mtart
          t_data-maktx
          t_data-meins
          t_data-matkl
          t_data-bismt
          t_data-extwg
          t_data-spart
          t_data-prdha
          t_data-MTPOS_MARA.
    append t_data.
    else.
    exit.
    endif.
    enddo.
    loop at t_data.
    *if sy-subrc <> 0. exit. endif.
    perform bdc_dynpro      using 'SAPLMGMM' '0060'.
    perform bdc_field       using 'BDC_CURSOR'
                                 'RMMG1-MATNR'.
    perform bdc_field       using 'BDC_OKCODE'
                                 '=AUSW'.
    perform bdc_field       using 'RMMG1-MATNR'
                                 t_data-MATNR.
    perform bdc_field       using 'RMMG1-MBRSH'
                                 t_data-MBRSH.
    perform bdc_field       using 'RMMG1-MTART'
                                 t_data-MTART.
    perform bdc_dynpro      using 'SAPLMGMM' '0070'.
    perform bdc_field       using 'BDC_CURSOR'
                                 'MSICHTAUSW-DYTXT(01)'.
    perform bdc_field       using 'BDC_OKCODE'
                                 '=ENTR'.
    perform bdc_field       using 'MSICHTAUSW-KZSEL(01)'
                                 'X'.
    perform bdc_dynpro      using 'SAPLMGMM' '4004'.
    perform bdc_field       using 'BDC_OKCODE'
                                 '/00'.
    perform bdc_field       using 'MAKT-MAKTX'
                                 t_data-MAKTX.
    perform bdc_field       using 'BDC_CURSOR'
                                 'MARA-PRDHA'.
    perform bdc_field       using 'MARA-MEINS'
                                 t_data-MEINS.
    perform bdc_field       using 'MARA-MATKL'
                                 t_data-MATKL.
    perform bdc_field       using 'MARA-BISMT'
                                 t_data-BISMT.
    perform bdc_field       using 'MARA-EXTWG'
                                 t_data-EXTWG.
    perform bdc_field       using 'MARA-SPART'
                                 t_data-SPART.
    perform bdc_field       using 'MARA-PRDHA'
                                 t_data-PRDHA.
    perform bdc_field       using 'MARA-MTPOS_MARA'
                                 t_data-MTPOS_MARA.
    perform bdc_dynpro      using 'SAPLSPO1' '0300'.
    perform bdc_field       using 'BDC_OKCODE'
                                 '=YES'.
    perform bdc_transaction using 'MM01'.
    endloop.
    *enddo.
    perform close_group.
    perform close_dataset using dataset.

  • Executing query in background : EXPORT_NO_SHARED_MEMORY

    Hi,
    I am using reporting agent to execute the query in background.The actual requirment is to print the report and send the report as mail after the triggeting of an event by a ABAP program.We want the same report to run with 9 different varints at the of trigger of this event.
    Let me explain what I have done.
    1.I went to reporting agent-->print
    2.Created 9 settings for the report
    3.Added all these settings in a package
    4.scheduled the package to run after the triggering of the event
    Now the job that executes the report in backgroung gets cancelled with error
    "ABAP/4 processor: EXPORT_NO_SHARED_MEMORY"
    When I checked the spool ,it showed result of the report with one variant.
    Is it because I have assigned the alll settinmgs under a single package ?
    Also can some one tell me about mailing the output  to the user?

    Hi Fazil,
    It sounds like a system problem. You might want to ask Basis people to help you. You can check trx ST22 or SM21 to get any log..
    Hope this helps.
    Regards,
    Diego

  • Dynamically Set date not changing in schedule Job

    Hi,
    I have created a report with "Order Date" as one of the parameters on the selection screen. I have created a variant, say 'X' , to set this date parameter dynamiccaly as "Current date - 31 days".
    I have scheduled this program to run in background with the 'X' variant daily at certain time.
    Now, for the day on which I created this job , the Order Date is set to correct date . For ex if I am scheduling the job tdy it will be set to 06/10/08. However, for next day , that is tom this date is not getting changed. i.e Tommorrow again Order date is 06/10/08 but it should have been 07/10/08.
    So basically this dynamically set date is not chaning in the scheduled job.
    Any idea why this is happening and what is the corrective measure for this.
    Thanks!

    There is an INITIALIZATION event in the program.......
    Do the processing your date in that event only....
    for ex....
    select-options: date for order-date.
    INITIALIZATION.
    date-low = sy-datum or wat ever.
    Thanks
    Saurabh

  • Background schedule process

    Hi
    I am new to Oracle 9iAS. Can someone tell me if there is any
    better way to implement a background scheduled process ?
    Currently, I am doing it via AT command. Just wondering if
    Oracle9iAS has any scheduler that I can make use of.
    Thanks
    CT

    You can set up the different step while schedule the back ground job. Each step will have the program name, variant.
    You can do his using transaction SM36 -> JOB WIZARD.
    Thanks,
    Srinivas

  • How to stop a Scheduler Job in Oracle BI Publisher 10g

    Hello!
    Can someone tell me how can I stop a scheduler job in Oracle BI Publisher 10g?
    I scheduled a bursting job to run a report but is running during two days.
    I would like to stop it.
    Thanks.
    Edited by: SFONS on 19-Jan-2012 07:16

    Unfortunately there is no way to stop a job once it is being executed. Yes as you read, it is not possible once job has started.
    Same thing applies for running queries.
    Once queries are sent to the DB BIP loses control over them. The message you see (if any) "Click Here to Cancel" does not stop any query
    it is just a message.
    I guess you will have to stop/kill the process in your DB
    regards
    Jorge
    p.s If you consider your question answered then please mark my answer as *"Correct"* or *"Helpful"*

  • OIM11g: Error while running AD Reconcilliation Schedule Job

    Hi All,
    I am using OIM 11g BP05 and have deployed AD connector 11.1.1.5.0 using connector server. While running the AD Group/Organization/User Reconcilliation schedule job, I am getting below error in connector logs:
    11/8/2012 4:25:59 PM <ERROR>: Class-> ActiveDirectoryUtils Method -> GetDirectoryEntry, Message -> Exception Stack Trace : at System.DirectoryServices.DirectoryEntry.Bind(Boolean throwIfFail)
    at System.DirectoryServices.DirectoryEntry.Bind()
    at System.DirectoryServices.DirectoryEntry.get_NativeObject()
    at Org.IdentityConnectors.ActiveDirectory.ActiveDirectoryUtils.GetDirectoryEntry(String path, ActiveDirectoryConfiguration configuration) in c:\ADE\aime_oimcp\idc\bundles\dotnet\ActiveDirectory\ActiveDirectoryConnector\ActiveDirectoryUtils.cs:line 1423
    11/8/2012 4:25:59 PM <ERROR>: Class-> ActiveDirectoryConnector Method -> ExecuteQuery, Message -> Caught Exception: Unable to get the Directory Entry11/8/2012 4:25:59 PM <ERROR>: Class-> ActiveDirectoryConnector Method -> ExecuteQuery, Message -> Stack Trace: at Org.IdentityConnectors.ActiveDirectory.ActiveDirectoryUtils.GetDirectoryEntry(String path, ActiveDirectoryConfiguration configuration) in c:\ADE\aime_oimcp\idc\bundles\dotnet\ActiveDirectory\ActiveDirectoryConnector\ActiveDirectoryUtils.cs:line 1456
    at Org.IdentityConnectors.ActiveDirectory.ActiveDirectoryConnector.ExecuteQuery(ObjectClass oclass, String query, ResultsHandler handler, OperationOptions options, Boolean includeDeleted, SortOption sortOption, String serverName, String searchRoot, SearchScope searchScope, OperationOptions otherOptions) in c:\ADE\aime_oimcp\idc\bundles\dotnet\ActiveDirectory\ActiveDirectoryConnector\ActiveDirectoryConnector.cs:line 1048
    at Org.IdentityConnectors.ActiveDirectory.ActiveDirectoryConnector.ExecuteQuery(ObjectClass oclass, String query, ResultsHandler handler, OperationOptions options) in c:\ADE\aime_oimcp\idc\bundles\dotnet\ActiveDirectory\ActiveDirectoryConnector\ActiveDirectoryConnector.cs:line 814
    ConnectorServer.exe Error: 0 : Org.IdentityConnectors.Framework.Common.Exceptions.ConnectorException: Unable to get the Directory Entry
    Also, I am seeing this error in log too:
    *11/8/2012 4:25:59 PM <ERROR>: Class-> ActiveDirectoryUtils Method -> GetDirectoryEntry, Message -> Exception occured during the creation of directory entry.*
    *11/8/2012 4:25:59 PM <ERROR>: Class-> ActiveDirectoryUtils Method -> GetDirectoryEntry, Message -> Exception Message : An invalid dn syntax has been specified.*
    11/8/2012 4:25:59 PM <ERROR>: Class-> ActiveDirectoryUtils Method -> GetDirectoryEntry, Message -> Exception Stack Trace : at System.DirectoryServices.DirectoryEntry.Bind(Boolean throwIfFail)
    at System.DirectoryServices.DirectoryEntry.Bind()
    at System.DirectoryServices.DirectoryEntry.get_NativeObject()
    at Org.IdentityConnectors.ActiveDirectory.ActiveDirectoryUtils.GetDirectoryEntry(String path, ActiveDirectoryConfiguration configuration) in
    Any pointers for resolving this issue?
    Regards,
    Sunny
    Edited by: ajmerasunny on Nov 8, 2012 4:32 PM

    Hi All,
    My bad, I gave wrong search base. Issue resolved.
    However, I need to run the a trusted source recon for only active
    Users. I gave filter as useraccountcontrol=512 but got casting exception.
    Any pointers on how to specify the filter to get only active accounts from AD?
    Regards,
    Sunny

  • Scheduled Job to gather stats for multiple tables - Oracle 11.2.0.1.0

    Hi,
    My Oracle DB Version is:
    BANNER Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    PL/SQL Release 11.2.0.1.0 - Production
    CORE 11.2.0.1.0 Production
    TNS for Linux: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - Production
    In our application, we have users uploading files resulting in insert of records into a table. file could contain records ranging from 10000 to 1 million records.
    I have written a procedure to bulk insert these records into this table using limit clause. After the insert, i noticed my queries run slow against these tables if huge files are uploaded simultaneously. After gathering stats, the cost reduces and the queries executed faster.
    We have 2 such tables which grow based on user file uploads. I would like to schedule a job to gather stats during a non peak hour apart from the nightly automated oracle job for these two tables.
    Is there a better way to do this?
    I plan to execute the below procedure as a scheduled job using DBMS_SCHEDULER.
    --Procedure
    create or replace
    PROCEDURE p_manual_gather_table_stats AS
    TYPE ttab
    IS
        TABLE OF VARCHAR2(30) INDEX BY PLS_INTEGER;
        ltab ttab;
    BEGIN
        ltab(1) := 'TAB1';
        ltab(2) := 'TAB2';
        FOR i IN ltab.first .. ltab.last
        LOOP
            dbms_stats.gather_table_stats(ownname => USER, tabname => ltab(i) , estimate_percent => dbms_stats.auto_sample_size,
            method_opt => 'for all indexed columns size auto', degree =>
            dbms_stats.auto_degree ,CASCADE => TRUE );
        END LOOP;
    END p_manual_gather_table_stats;
    --Scheduled Job
    BEGIN
        -- Job defined entirely by the CREATE JOB procedure.
        DBMS_SCHEDULER.create_job ( job_name => 'MANUAL_GATHER_TABLE_STATS',
        job_type => 'PLSQL_BLOCK',
        job_action => 'BEGIN p_manual_gather_table_stats; END;',
        start_date => SYSTIMESTAMP,
        repeat_interval => 'FREQ=DAILY; BYHOUR=12;BYMINUTE=45;BYSECOND=0',
        end_date => NULL,
        enabled => TRUE,
        comments => 'Job to manually gather stats for tables: TAB1,TAB2. Runs at 12:45 Daily.');
    END;Thanks,
    Somiya

    The question was, is there a better way, and you partly answered it.
    Somiya, you have to be sure the queries have appropriate statistics when the queries are being run. In addition, if the queries are being run while data is being loaded, that is going to slow things down regardless, for several possible reasons, such as resource contention, inappropriate statistics, and having to maintain a read consistent view for each query.
    The default collection job decides for each table based on changes it perceives in the data. You probably don't want the default collection job to deal with those tables. You probably do want to do what Dan suggested with the statistics. But it's hard to tell from your description. Is the data volume and distribution volatile? You surely want representative statistics available when each query is started. You may want to use all the plan stability features available to tell the optimizer to do the right thing (see for example http://jonathanlewis.wordpress.com/2011/01/12/fake-baselines/ ). You may want to just give up and use dynamic sampling, I don't know, entire books, blogs and papers have been written on the subject. It's sufficiently advanced technology to appear as magic.

  • Background Scheduling - Deadline Monitoring

    Hi Experts,
    I have created new maintenance plan , its around 250 for Mechanical and 250 for Electrical
    Maintenance plan having strategy like every two month , every three month , every six month etc..
    Scheduling period is 3 year
    Now scheduling will be done through IP10 one time , but after first call object scheduling will get hold.
    I want to make it background scheduling - Deadline Monitoring
    Now the requirement is , For Example :
    Maintenance plan X have PM plan date is 21.07.2014
    Maintenance plan Y have PM plan date is 24.07.2014
    Maintenance plan Z have PM plan date is 29.07.2014
    As per above example each equipment will have different plan date , background scheduling can be done through IP30 but by the date is concern how can make it possible ?
    Please advise each maintenance plan has to be run as per their plan date.
    Thanks,
    Abhishek

    Hi Mr.Jogeshwara,
    Nice Document , Its really useful
    But my query is each equipment have different plan date , as explained in above example
    If Maintenance plan have Cycle Unit 60 day and Plan X date is on 21.07.2014 then 21.07.2014+60day , same as if maintenance plan Y date is 25.07.2014 then 25.07.2014 +60day
    Like that
    background scheduling needs to be done as per cycle unit
    If i will do background scheduling together of both the above maintenance plan X and Y then scheduling will be run exactly after 60 day and order will be generated in advance for Maintenance plan Y.
    Hope you understand my query.
    Thanks,
    Abhishek

  • Background Scheduling Z Report

    Hi,
    I have developed one report , it can output its result in alv report and excel file, So my requirement is to schedule the report and in background only email the report to the specified email id's. How can I achieve functionality of background scheduling and automatic email.
    I know we can do background scheduling with SA38 or SM36/SM37 but upon scheduling it is not saving my excel file at desired location.
    Ankesh

    You should search forums about this....
    You will need to save to the SAP Application server (in a folder which has write permission for SAP jobs) and possibly FTP to some shared drive within the network.  Local PC, local MS network, like C:\, D:\, etc. is not visible to a background job, since there is NO GUI/PRESENTATION server involved in a background job.  Outside of MS Office, an .XLS file is an unknown file type, probably....  how do you think you're creating a .XLS file in a background SAP job?

Maybe you are looking for

  • How do you retrieve a Hard Drive on a mac OSX?

    ok, here goes recently my Mac OSX froze so i immediately turned it off , so when i turned it back on the loading screen had a file with a Question mark on it, then i went to my room to get my Mac OSX installer Disc and i put in disc one and waited fo

  • Indesign Cs6 won't export to pdf

    I am publishing  a book and every time I try to upload it to a pdf it says: "Cannot save to file because it is in use by one or more background tasks. I have read a few other forums but I could not figure out how to fix mine. It is adobe Indesign cs6

  • Using an AIX archive file as a library for an external procedure.

    I cannot use an archive file, created with the ar command, that contains a shared object that was created with the xlc and ld commands as a library for external procedures. I can use the shared object file created with the ld command as a library for

  • Web developer tolbar

    Hello Experts, I am so grateful for all the excellent support I get on this forum. I am following CSS tutorials on Lynda.com, in which they use the "Web Developers Tool Bar". It is a Dreamweaver extension, I believe, that when implemented establishes

  • Restrict users from changing password on first login?

    Hi, I am doing mass user upload into UME using script import. How should I use the below functionality to restrict the users from changing password on first login? IUserAccount uacc =UMFactory.getUserAccountFactory().newUserAccount(uid,newUser.getUni