Scheduling job perodically vs schedule manager

Dear SAP expert
I would like to run the report fbl1n(PRogram  RFITEMAP ) and email to users.
I have 2 option:-
1.Schedule the job periodically and email it through SM36.
2. customise the report in schedule manager in flow definition and email to users.
option 1 i am able to execute.
option 2 i am not fully through. Please clarify how option 2 can be done.
How i should decide pros and cons of schedule the job periodically through SM36 OR THROUGH SCHEDULE MANAGER
PLS CLARIFY
thanks
S AP

Hello Santosh,
the link that you already got is good. Perhaps you also want to check out [http://wiki.sdn.sap.com/wiki/display/SM/FAQJobSchedulingManagement|http://wiki.sdn.sap.com/wiki/display/SM/FAQJobSchedulingManagement] and [SAP is first vendor to provide a comprehensive Job Management Suite |SAP is first vendor to provide a comprehensive Job Management Suite] for further information.
In short SAP CPS is an advanced scheduling tool like which competes against other like UC4, Control-M from BMC etc. But all these tools do not provide any real job documentation or job request functionality in order to get the process governance right. That's where the SolMan gets into the game.
Best Regards
Volker

Similar Messages

  • ODI Scheduled Jobs

    Hi,
    I have more than 100 jobs scheduled in my work repository.Difficult to maintain a spreadsheet and keep track of scheduled jobs and their scheduled timings.
    Are there any database tables at back end which has list of all scheduled jobs and their corresponding scheduled timings?
    Any help on this would be really appreciated.
    Thanks and Regards
    Reshma

    Hi 2755523,
    You can query the repository :
    SELECT * FROM SNP_PLAN_AGENT;
    You can find the repository structure in Doc ID 1903225.1 on My Oracle Support : https://support.oracle.com/epmos/faces/DocumentDisplay?_afrLoop=545729001970661&id=1903225.1&_adf.ctrl-state=8ftus7sby_5…
    Regards,
    JeromeFr

  • Manually kick off scheduled jobs

    Hi all, now and again I have a situation where the database that the BIP reports are run against is not availalbe in the morning when the BIP jobs are scheduled to run. When the database comes online mid-morning, I've been manually running all jobs that were scheduled to run earlier that morning.
    Is there a way to manually kick off the scheduled jobs? The scheduled job already contains the report output name, format, destination etc. and having to go back to the original reports and manually run takes a lot of time. Any ideas?

    Hi Joe,
    I have done migration Discoverer Admin EUL Layer into OBIEE repository using below methodology.
    Navigate to the <installdrive>\OracleBI\server\Bin directory. There are two important files in this directory: the migration assistant executable file named MigrateEUL.exe and a properties configuration file named MigrationConfig.properties.
    Could you please help me how to migrate discoverer plus workbooks and worksheets into OBIEE Answers?
    go through below link, It will show navigation steps for migrating of EUL from Discoverer to OBIEE.But i need migration of workbooks and worksheets from Discoverer into OBIEE Answers.
    http://www.oracle.com/technology/obe/obe_bi/discoverer/discoverer_1012/discomigration/migrate_disco_biee.htm
    This is very great full help to me …
    Advance thanks for your suggestions.
    Regards
    Duraga Prasad.

  • Error while running scheduled job fromEnterprise Manager--Please help ASAP

    BEGIN
    Hi,
    I have created a scheduled job using the below code.
    BEGIN
    sys.dbms_scheduler.create_job(
    job_name => '"M2RT6"."INCR_REFRESH_M2RT6"',
    job_type => 'PLSQL_BLOCK',
    job_action => 'begin
    REFRESH_REPORTS_POETIC(''M2PT6'',''INCR'');
    end;',
    start_date => systimestamp at time zone 'US/Eastern',
    job_class => 'DEFAULT_JOB_CLASS',
    auto_drop => FALSE,
    enabled => TRUE);
    END;
    And I tried to execute this job from Enterprise Manager.
    I got the below error:
    java.sql.SQLException:
    ORA-02800: Requests timed out
    ORA-06512: at "SYS.DBMS_ISCHED", line 150
    ORA-06512: at "SYS.DBMS_SCHEDULER", line 441
    ORA-06512: at line 2
    Help needed urgently. Thanks for the resolution inadvance.
    Thanks,
    Sudheer.Vempali

    Hi,
    This is the bug in Oracle. Most of the times, restarting Oracle database may fix the issue.
    Basically, an existing bug needs the oracle database restarts in a regular basis.
    Restarting oracle database will flushes the database buffer cache.
    Thanks,
    Sudheer.Vempali

  • Report Queue Manager  - Scheduled Jobs

    Question: How/where does the Queue Manager store info for scheduled jobs?

    It is possible (IN 9iAS) to configure the Reports Server to store all requests in a database table. The Table(s) / View(s) are created by the script rw_server.sql located in $ORACLE_HOME/reports/admin/sql
    To configure this options, create the table and view with the rw_server.sql script, then add these lines in the
    <reports server name>.conf file <jobStatusRepository class="oracle.reports.server.JobRepositoryDB"> <property name="repositoryConn" value="user/password@tnsalias" confidential="yes" encrypted="no"/> </jobStatusRepository> The "value" attribute of the property "repositoryConn" should be a user/password@tnsalias that will be used to connect to the database. The account must have access to the tables created by the script rw_server.sql The combination confidential="yes" and encrypted="no" will provoke the encryption of the value attribute once the Report Server is re-started : After the Reports Server has started you will find: <jobStatusRepository class="oracle.reports.server.JobRepositoryDB"> <property name="repositoryConn" value="ihjFFBFD3jk5Xfrd5HeX5w3dE/E=" confidential="yes" encrypted="yes"/> </jobStatusRepository> Description of the rw_server_queue view columns can be found in the script rw_server.sql itself.
    Thanks,

  • BO4 - Lifecycel Management Console to export all content via scheduled job

    Hi,
    We wish to export the entire content of our repository to a lcmbiar file via scheduled job every week.
    Whilst could do this manually, want this to be a periodic refresh.
    The purpose of this is to have a developemnt system which has the content of the production system as of a week ago.
    So would want this job to be dynamic, exporting entire content including any new content on weekly basis at week-end during quiet time on servers.
    Anybody know if this can be done via lcmbiar file automatically?
    I'm thinking say sales folder week 1 10 reports, week 2 somebody creates another 2 reports, would want lcmbiar file to automatically pick up the new reports which have been added in week 2.
    Is this possible via lifecycle management console?
    Can't simply export cms database to development system as contains the production host embedded within it and don't want the development system to have reference to the physical production host.
    Many Thnaks

    Hi,
    Yesterday ASUG conducted webinar on LCM 4.0 FP3. URL for webbex session
    http://www.asug.com/Default.aspx?tabid=124&vp_url=http://jive.asug.com/docs/DOC-32206
    summary:
    1. We can not schedule to export LCMBIAR file in 4.0 SP2
    2. When ant file added to folder, automatically scheduled job do not pick the file in 4.0 SP2
    But above two options are good and working in FP3 as per SAP.

  • Oc4j scheduled jobs management

    We have an application deployed under oc4j server and that launches scheduled jobs.
    When we restart the oc4j server, old jobs are relaunched automatically which generate exception. Is this behavior logical? How we can remove these jobs.
    Thanks for collaboration

    Hi,
    Some diagnostic things to check
    - are there any crashes reported in the alert.log ?
    - are any jobs running
    select * from dba_scheduler_running_jobs;
    - is the job coordinator running
    on UNIX
    ps aux |grep cjq
    select program from v$session; -- look for CJQ
    - are job slaves running
    on UNIX
    ps aux |grep j0
    select program from v$session; -- look for J0*
    - on UNIX what process is consuming the CPU (check output of top and then grep the output of "ps aux" for the process ID)
    - what is MAX_JOB_SLAVE_PROCESSES set to ?
    select * from all_scheduler_global_attribute where ATTRIBUTE_NAME='MAX_JOB_SLAVE_PROCESSES';
    - what is job_queue_processes set to ?
    select value from v$parameter where lower(name)='job_queue_processes';
    - what is the status of all enabled jobs
    select job_name, state, next_run_date, last_start_date from dba_scheduler_jobs
    where state!='DISABLED';
    If scheduled jobs are not running and there are no jobs running, there is really nothing to "reset", it's just a matter of figuring out why jobs are not running. If you provide answers to the questions above, it should become clearer what is happening.
    One thing you can try is disabling all jobs and reenabling them one by one to see if they run.
    The purge_log job running more than a day is extremely unusual, but it shouldn't step on itself as it will just skip the next run if it is already running.
    Hope this helps,
    Ravi.

  • Oracle scheduler : Setting job attributes using Enterprise Manager

    Hi,
    I have created below job and set the attribute "'raise_events" using the dbms_scheduler.set_attribute procedure.
    begin
    dbms_scheduler.create_job
    ( 'ebj_A', job_action =>
    '/tmp/job_A.ksh',
    job_type => 'Executable',
    enabled => false ,
    auto_drop =>false) ;
    dbms_scheduler.set_attribute
    ( 'ebj_A' , 'raise_events' , dbms_scheduler.job_run_completed);
    end;
    Now I am creating the same job using the Enterprise Manager (GUI).
    I donot see any where to set the attribute (Raise_events) using GUI.
    Please can you tell me can we set that attribute using the Enterprise manager or not.
    If yes please let me know how.
    Thanks in advance.
    Sreenath.

    Hi ,
    I am modified my question.
    We could set the value for "raise_events" attribute using Dbms_Scheduler.Set_Attribute.
    Can we also set the value for the "raise_events" attribute using the EM.(GUI).
    Please let me know.
    Thanks in advance.
    Sreenath.

  • ADF_FACES-60097 error while opening a random schedule job

    Hello,
    recently i'm facing a error in oim while i open the schedule job
    When i logeed in with xelsysadm
    click on advanced.then click on System management
    search all the scehdule jobs
    open a random schedule job
    error occured!!
    Error: An error Occurred while retrieving trigger for the schedule job.
    ADF_FACES-60097
    please help
    regards
    tp

    Hi Frank,
    Thanks for the reply. I am using it as an inline popup. I am trying to create command links dynamically in an iterator. When a link is clicked a popup window opens. The text of the link is passed to the popup window by creating a variable in the session scope ( I even tried request scope) by an action listener in the backing bean.
    I tried to create another simple application with a static command button which passes its text to popup window through a session scope variable. It worked fine. I am not sure why the same procedure is not working in my original application. Can you give me a hint what could possibly go wrong according the exception stack trace ?
    Thanks,
    KK

  • How to schedule job V3 unserialised Update job for Inventory

    Hi gurus,
    How to schedule job V3 unserialised Update job for Inventory management
    What are the things need to be configured
    Thank you

    Hi,
    In R/3 go to SBIW > Settings for Application-Specific DataSources (PI) > Logistics > Managing Extract Structures > Logistics Extraction Structures Customizing Cockpit
    UIn this Screen you can define delta type, and the jog usign the control job option.
    Regards,
    Dani

  • Regarding error while scheduling Job Server (rapid mart)

    hi friends
    iam new to bods we are using BODS 3.2 source r/3 and target is MYSQLSERVER . and we have installed rapidmarts
    General_ledger_rapid mart during scheduling the job fails and got and
    erorr
    Transform EID_2_Rename_ConvertDateToChar Cannot convert string <1900.01.01 00:00:00> to date using format string <YYYY-MM-DD>.
    (12.2) 02-11-12 04:59:09 (W) (1124:11052) RUN-053006: |Data flow DF_FinDocFactNewGL_SAP|Embedded Dataflow Copy_1_EDF_FinDocFact_SAP|Transform EID_2_Rename_ConvertDateToChar
    please help how to rectify this step by step ( as iam new )
    please

    thanks suneer
    can you please explain if i find the problem and rectify
    and after that how could i schedule the jobs ( as it is a rapid mart it contains number of data flow) so,
    how should i schedule or do i need to schedule this DF only in the designer or through Data Service Management console
    I think if i schedule it will again schedule all the jobs in the rapid mart ( i want to schedule only one DF which is eror so, how should i do)
    and if it is"? it will show duplicate reports
    please suggest
    thanks

  • Problem with variable in scheduled job

    I'm trying to get the following scheduled job to run:
    switch(config)# scheduler job name backup_job
    switch(config-job)# cli var name timestamp $(TIMESTAMP) ; copy running-config bootflash:/$(SWITCHNAME)-cfg.$(timestamp) ; copy bootflash:/$(SWITCHNAME)-cfg.$(timestamp) tftp://1.2.3.4/
    switch(config-job)# exit
    switch(config)# scheduler schedule name backup_timetable
    switch(config-schedule)# job name backup_job
    switch(config-schedule)# time daily 1:23
    switch(config-schedule)# exit
    switch(config)# exit
    This job is taken directly from multiple Cisco MDS and Nexus documents. From what I can tell, the purpose of this job is to save the running configuration to a file on bootflash with date & time in the file name and then to copy the file from bootflash to tftp server.
    I can create the job and schedule successfully:
    switch(config)#show scheduler job name backup_job
    Job Name: backup_job
    cli var name timestamp $(TIMESTAMP)
    copy running-config bootflash:/$(SWITCHNAME)-cfg.$(timestamp)
    copy bootflash:/$(SWITCHNAME)-cfg.$(timestamp) tftp://1.2.3.4
    ==============================================================================
    switch(config)#show scheduler schedule name backup_timetable
    Schedule Name       : backup_timetable
    User Name           : admin
    Schedule Type       : Run every day at 10 Hrs 48 Mins
    Last Execution Time : Tue Mar  6 10:48:00 2012
    Last Completion Time: Tue Mar  6 10:48:00 2012
    Execution count     : 1
         Job Name            Last Execution Status
    backup_job                        Success (0)
    ==============================================================================
    The scheduled job runs successfully but the files that are created have the variable $(TIMESTAMP) in the file name instead of the actual date and time e.g. switch-cfg.$(TIMESTAMP)
    The logfile contains the following:
    Schedule Name  : backup_timetable                  User Name : admin
    Completion time: Tue Mar  6 10:59:26 2012
    --------------------------------- Job Output ---------------------------------
    `cli var name timestamp $(TIMESTAMP)`
    `copy running-config bootflash:/PEN-9509-2-cfg.$(TIMESTAMP) `
    Copy complete, now saving to disk (please wait)...
    `copy bootflash:/PEN-9509-2-cfg.$(TIMESTAMP) tftp://1.2.3.4 `
    Trying to connect to tftp server......
    Connection to server Established. Copying Started.....
    It looks to me that the $(timestamp) variable is being created successfully and is being replaced with the $(TIMESTAMP) variable but this is not being replaced with the actual date and time.
    The thing I don't get is that this looks to me that we're trying to nest variables and the same Cisco documents from which I get this configuration also state that nested variables are not allowed.
    I have tried this on different hardware - MDS9500, MDS9100, Nexus 5000, Nexus 7000 and different software - SAN-OS 3.3, NX-OS 4.1, NX-OS 5.2 but cannot get it to work. I have also tried to put the commands in a script and run with the run-script command but it still does not work.
    There is probably another method to achieve what this configuration is trying to achieve (and I would like to know if there is) but I want to know if this particular configuration will work.
    Can anyone tell me if they have got this working or can see what I'm doing wrong or can try running this in a lab please?

    I managed to get this resolved with a bit of a workaround. If I put the copy commands in a script and pass the variable to the run-script command as part of the scheduled job then it works ok. Trying to create the variable within the script (or as a separate scheduled job command) still doesn’t work.
    So, creating a script file (script) as follows:
    copy running-config bootflash:/$(SWITCHNAME)-cfg.$(timestamp)
    copy bootflash:/$(SWITCHNAME)-cfg.$(timestamp) tftp://1.2.3.4
    and creating a scheduled job with the following command:
    run-script bootflash:script timestamp=”$(TIMESTAMP)”
    achieves the desired result.

  • Event based scheduler job - 2 events at the same time only 1 run

    Hi,
    i converted our dbms_job - jobs to the newer package dbms_scheduler.
    It is a 10.2.0.4 patch 23(8609347) database on an windows server 2003 R2 Enterprise x64 Edition SP2.
    The Jobs(about 130) are nothing special ... only some statistics, matview-refreshes and so on.
    For the notification of failed jobs and jobs which run over the max_run_duration i downloaded and installed the job notification package.
    The jobs are assigned to different departments and the corresponding developer teams in our company.
    I created a notification job for each department and if a job fails we (the database administrators) and the corresponding deverlopers will be informed.
    Now i ascertained that only 1 email will be send if 2 jobs of the same department fails at the same time.
    The emailer-jobs are auto-generated by the job notification package. I only modified them to look after all jobs of special department and not only after one job. (--> event_condition ... object_name LIKE 'XXX%')
    example for dba-jobs(copy of the script output of TOAD):
    SYS.DBMS_SCHEDULER.CREATE_JOB
           job_name        => 'DBA_JOBS_EMAILER'
          ,start_date      => NULL
          ,event_condition => tab.user_data.object_name LIKE ''DBA%'' AND tab.user_data.event_type in (''JOB_FAILED'',''JOB_OVER_MAX_DUR'')'
          ,queue_spec      => 'SYS.SCHEDULER$_EVENT_QUEUE, JOB_FAILED_AGENT'
          ,end_date        => NULL
          ,program_name    => 'SYS.EMAIL_NOTIFICATION_PROGRAM'
          ,comments        => 'Auto-generated job to send email alerts for jobs "DBA%"'
        );I thought that a queue is used to manage all events from the scheduler jobs. So i made a test with 2 dba jobs and simulated a failure at the same time but i received only one mail.
    So what is happend with the second event? I looked for the events in the qtab(SCHEDULER$_EVENT_QTAB) which belongs to the event queue(SYS.SCHEDULER$_EVENT_QUEUE) and no event was missing.
    So i think the emailer job has to run 2 times.
    Is anyone able to explain or to find my mistake?
    I know that the easiest way is to create one emailer job for each normal job but i think this is a little bit costly because all the arguments are the same for one department.
    Thanks & Regards

    Thanks for your fast answer.
    You are right with the "enabled => TRUE;" part and i only forgot to post it.
    So the Job is enabled (otherwise it would not send any mail). Because it is sending one mail i think it is also not necessary to hand over a job_type.
    Additionally the job starts a program ... so it is right to set the job_type='STORED_PROCEDURE' isn't it?
    And also right ... i already added the agent as subscriber to the queue.
    Anyway i think the whole thing do what it have to do. So in my oppinion there are no big mistakes in creating the job or at adding the subscriber.
    There are also no problem in raising the events by itself and enqueue them in the scheduler event queue.
    There is only a problem when 2 jobs fails (or run out ouf max duration) at exactly the same time.
    If i understand it right:
    The agent/subscriber will find the "JOB_FAILED"-event for the first Job in the queue and starts the emailer Job.
    The agent will also find the "JOB_FAILED"-event for the second Job and wants to start the emailer Job again.
    I don't know if this is really the problem but perhaps the emailer-job can not be started in consequence of the second event because ist is already running.
    I also don't know if this is a mistake of the agent or of the emailer-job by itself.
    I only want that it runs two times (one run for each event). I my case it also doesn't matter which email is send at first.

  • Basic Help needed with Hyperion 9.3 Scheduling jobs

    Hi!
    Is there anyone who would be willing to help a novice understand the Scheduling features in Hyperion 9.3? We are converting from Brio 6.6 where we used the "Broadcast Server". We don't yet have many jobs - so I'm not migrating them. But, the documentation with Hyperion is not helping me really understand how to set a job to actually do anything. I don't yet know how to add cycles and direct specific output to specific places. The jobs "run" and I've created events ...
    If you would have some time to talk me through the process, I would be very grateful.
    Jeanne Loehnis
    [email protected]
    920-832-6613

    I agree about the documentation. It's not always easy to understand how to do something if your situation doesn't exactly match the examples. I had to use a lot of trial and error to finally get some jobs running.
    The first misunderstanding I had was that if I want to schedule a document, I have to import my .bqy file as a job from the filesystem. You can't take an Interactive Reporting document from Workspace and use it as the basis of a job. So if you're using the web-based client and working on documents stored in Workspace you have to export the document to your filesystem and then import it as a job. You're probably beyond that point.
    The second misunderstanding (and I wonder if this is your problem) is that you must set the job to allow users to create their own cycles for it. If you don't you're stuck with the default cycle set, and you cannot change it. This is where the documentation falls short -- it mentions the attribute, but doesn't go far enough to tell you what happens if you don't check it off -- that even you, as the owner-and-most-likely-also-administrator won't even be able to maninpulate the cycles. What's more, if you don't set this attribute on import, you can't set it later. There's no way to reach the dialog with Properties. I can't vouch for the Batch Scheduler, we don't license that (if it even applies). Maybe it's a bug, maybe I'm missing something, but this is the only way I can get the full functionality to work.
    You will find this attribute on Step 3 of the import dialog "Interactive Reporting Job Properties". Check the box labeled "Users running this job can define their own cycles and actions". Once you finish and schedule the job, you can go to Manage Job Schedules and modify the schedule properties. Go to the Select Job Parameters tab, and create or use an existing parameter set (using the Go button).
    This takes you to the Job Parameter dialog box, and that is where you set up your cycles and actions. You can set them up, save them as your own personal sets, or as public sets so your team can use them for other jobs. You should be able to make some headway here, the documentation becomes more useful again at this point.
    Hope this helps
    --Julie
    Julie Johnson
    University Computing
    Western Connecticut State University

  • DPM 2010 Scheduled Jobs Disappear rather than Run

    I have a situation where I have a DPM server that appears to be functioning fine, but none of the scheduled jobs run.  No errors are given, there are no Alerts, and there is nothing in the Event log (Application and System) which indicates a failure. 
    All my Protection Groups show a green tick to indicate that they are fine, but the last successful backup for all of them is Friday the 8th of February.
    If I go to Monitoring and Jobs I see the jobs scheduled, but when the time comes for the job to run, it does not go into the "All jobs in progress", it just merely disappears, like thus:
    And a few minutes later,
    As you can see, the jobs disappear from the queue, and the total number of jobs decreases accordingly.  These jobs do not go into any of the other 3 Statusses (Completed, Failed or In Progress), they just disappear without a trace.
    There is some unallocated space, albeit not much (Used space: 21 155,05 GB Unallocated space: 469,16 GB). If space was an issue I would expect to see errors to indicate this.
    DPM 2010 running version 3.0.8193.0 (hotfix rollup package 6) using remote instance of SQL 2008 which is functioning fine.  I have tried stopping/starting the services, and even rebooted the server twice.  The remote instance of SQL server is using
    a domain account as its service account.  There are no pending Windows updates, i.e. it is fully up-to-date.
    The System Center Data Protection Manager 2010 Troubleshooting Guide (July 2010) does not show how to troubleshoot this particular probelm.
    Does anybody know how to resolve this issue or which logs might help me troubleshoot it?

    OK,
    Did you change the SQL Agent user account ?
    If so, DPM enters the SQL Agent account name into the registry and later we check that account each time the DPM engine launches.  The internal interfaces to DPM are secured using this account so the account name needs to match the account the SQL Agent
    is using. 
    Step 1
    In the registry HKLM\Software\Microsoft\Microsoft Data Protection Manager\Setup alter  both
    SqlAgentAccountName and SchedulerJobOwnerName keys to reflect the SQL Agent user account being used.
    Step 2
    Update DCOM launch and access permissions to match what was granted to the Microsoft$DPM$Acct account.
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread. Regards, Mike J. [MSFT]
    This posting is provided "AS IS" with no warranties, and confers no rights.

Maybe you are looking for