Is it possible to parameterize a scheduled job?

I have the following scenario which I consider implementing using Azure scheduler:
- Request information from an external Web service
- Store the result in the Azure blob
When requesting the information I need to supply an HTTP header indicating since what time to retrieve the requested information. So with each new request I need to update such header.
I checked Azure scheduler but its configuration seems to be static, i.e. I have to predefine all request parameters. Is it so?
If it is so, then what would be the most practical way of implementing such schedule? Should I build a simple intermediate Web service that will take care of request parameterisation so I can call it using static job configuration? It looks straightforward,
I just don't want to create middleware if it can be avoided.
Thanks in advance
Vagif Abilov

Hi Vagif,
Thanks for your posting!
I totally understand your requirement.  You could custom the headers using scheduler API. Please see this blog(
http://fabriccontroller.net/blog/posts/a-complete-overview-to-get-started-with-the-windows-azure-scheduler/ ) and code sample (https://github.com/bradygaster/maml-demo-scheduled-webjob-creator/blob/master/ScheduledWebJobCreator/Program.cs
Hope this helps.
Regards,
Will
We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
Click
HERE to participate the survey.

Similar Messages

  • Possibility of deleting the scheduled jobs through reports

    hi friends
    is it possible to delete the schedule jobs through report making?
    i mean i have to enter a particular user name
    list of jobs scheduled by him has to be listed
    and selecting the unwanted job have to delete all the jobs
    Thanks and regards
    s.janagar

    Jangar,
    Yes. This can be done. Write a delete statement and use it on the table TBTCO.
    Note: Any other operations can be done on the tables TBTC*
    Thanks,
    Babu Kilari

  • Is it possible to run scheduled job

    Is it possible to run scheduled job (custom developed java application), for doing some periodic database operstions in portal.
    I'm using portal database (maxdb) with my portal application.I must do some periodic calculations and store them into database table.Therefore i must use scheduled job running on ortal server.
    Is it possible? If possible, how?
    Thanks.

    Hi,
    Yes it is possible with SAP NetWeaver Scheduler for Java.
    You can check below link for more information:
    http://help.sap.com/saphelp_nwce71/helpdata/en/44/03d66015ee10b3e10000000a11466f/content.htm
    http://help.sap.com/saphelp_nwce71/helpdata/en/45/45d287e620507be10000000a1553f6/content.htm
    Thanks
    Sunny

  • ADF_FACES-60097 error while opening a random schedule job

    Hello,
    recently i'm facing a error in oim while i open the schedule job
    When i logeed in with xelsysadm
    click on advanced.then click on System management
    search all the scehdule jobs
    open a random schedule job
    error occured!!
    Error: An error Occurred while retrieving trigger for the schedule job.
    ADF_FACES-60097
    please help
    regards
    tp

    Hi Frank,
    Thanks for the reply. I am using it as an inline popup. I am trying to create command links dynamically in an iterator. When a link is clicked a popup window opens. The text of the link is passed to the popup window by creating a variable in the session scope ( I even tried request scope) by an action listener in the backing bean.
    I tried to create another simple application with a static command button which passes its text to popup window through a session scope variable. It worked fine. I am not sure why the same procedure is not working in my original application. Can you give me a hint what could possibly go wrong according the exception stack trace ?
    Thanks,
    KK

  • Background Processing? how schedule job for "System Error" Message .

    Hello everyone,
    in sap help i have read.
    http://help.sap.com/saphelp_nw04/helpdata/en/5a/f72040599a8f5ce10000000a155106/frameset.htm
    PCK> Monitoring>Message Monitoring-->Background Processing
    you can schedule jobs for various background processing:
    ●     Archiving of messages processed successfully
    ●     Deletion of messages that are not to be archived
    ●     Restarting of messages with errors
    ●     Rescheduling of lost messages
    can anyone understand this docu?
    give me some introduction, how can i define and schedule these jobs ?
    thx in advance!!
    best regards
    Yaning

    Background Processing
    Prerequisites
    You have started the message monitor on the initial screen of the PCK and are in Background Processing.
    Features
    Archiving
    You require two archiving sessions to archive messages:
    ●     One session to write the messages to the archive
    ●     One session to delete the persisted messages that have been archived
    To do this, you schedule an archiving job, which implicitly schedules the sessions to write to the archive and delete the archived messages.
    You can define one or more rules for each archiving job; these rules contain conditions that a message must meet in order to be archived by the job. At least one of the defined rules must be met for archiving to take place.
    All information that is displayed for a message in message monitoring is archived, in addition to the audit log for each message.
    Deleting
    A standard delete job is created automatically. It runs once a day. You can schedule additional delete jobs; however, you cannot define rules for them.
    Restarting
    Instead of restarting messages with errors manually with message monitoring, you can schedule a job to automatically restart these messages. This is possible for all messages for which the number of defined restart attempts has been exceeded (messages with the system error status).
    You can define one or more rules for each job to restart messages; these rules contain conditions that a message must meet in order to be restarted by the job. At least one of the defined rules must be met for archiving to take place.
    Rescheduling
    A standard job to reschedule messages is created automatically. The job runs once a day and ensures that messages lost as a result of database failure, for example, are rescheduled. You can schedule additional rescheduling jobs; however, you cannot define rules for them.
    Thx Aamir.
    But I mean the messages with errors in Adapter Engine , not in Intergrations Engine.
    the situation is like Naveen Pandrangi's WebLog
    II. Errors in Adapter Engine [XI :  How to Re-Process failed XI Messages Automatically|XI :  How to Re-Process failed XI Messages Automatically]
    I
    Till now we have seen how to resubmit/restart message that failed in Integration Engine.  One a message makes it from Integration Engine to Adapter Engine, the message is flagged as checked in Integration Engine. The status of the message in Adapter engine does not effect the processed state in Integration Engine. Now if this message was asynchronous, XI will by default try to restart the message 3 times at intervals of 5 minutes before the status of the message is changed from Waiting to System Error .
    *how can i schedule a job to automatically restart these messages with errors?
    best regards
    Yaning
    Edited by: Yaning Liu on Aug 18, 2008 1:43 PM

  • How to schedule job?

    Hello All,
    i have a few questions about jobs (scheduling jobs)
    I want to run one procedure at 7:00 A.M and I want to stop this procedure at 12:30 PM on Daily basis.Is it possible with DBMS_JOB? I don't know how to stop running job at specified time.
    I m using Oracle 9i on Linux

    Hi,
    Your best bet is probably to schedule a second job to run at 12:30pm every day and execute the stop. To do the stop you will probably need to figure out the session and terminate the session (google for stopping dbms_job).
    In 10gR2 and up dbms_scheduler (the replacement for dbms_job) can handle this scenario in a much more striahgtforward way (generate an event if the job exceeds a max_run_duration and have the event trigger a second job to snipe the first using a stop_job call).
    -Ravi

  • How to stop a Scheduler Job in Oracle BI Publisher 10g

    Hello!
    Can someone tell me how can I stop a scheduler job in Oracle BI Publisher 10g?
    I scheduled a bursting job to run a report but is running during two days.
    I would like to stop it.
    Thanks.
    Edited by: SFONS on 19-Jan-2012 07:16

    Unfortunately there is no way to stop a job once it is being executed. Yes as you read, it is not possible once job has started.
    Same thing applies for running queries.
    Once queries are sent to the DB BIP loses control over them. The message you see (if any) "Click Here to Cancel" does not stop any query
    it is just a message.
    I guess you will have to stop/kill the process in your DB
    regards
    Jorge
    p.s If you consider your question answered then please mark my answer as *"Correct"* or *"Helpful"*

  • Scheduled Job to gather stats for multiple tables - Oracle 11.2.0.1.0

    Hi,
    My Oracle DB Version is:
    BANNER Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    PL/SQL Release 11.2.0.1.0 - Production
    CORE 11.2.0.1.0 Production
    TNS for Linux: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - Production
    In our application, we have users uploading files resulting in insert of records into a table. file could contain records ranging from 10000 to 1 million records.
    I have written a procedure to bulk insert these records into this table using limit clause. After the insert, i noticed my queries run slow against these tables if huge files are uploaded simultaneously. After gathering stats, the cost reduces and the queries executed faster.
    We have 2 such tables which grow based on user file uploads. I would like to schedule a job to gather stats during a non peak hour apart from the nightly automated oracle job for these two tables.
    Is there a better way to do this?
    I plan to execute the below procedure as a scheduled job using DBMS_SCHEDULER.
    --Procedure
    create or replace
    PROCEDURE p_manual_gather_table_stats AS
    TYPE ttab
    IS
        TABLE OF VARCHAR2(30) INDEX BY PLS_INTEGER;
        ltab ttab;
    BEGIN
        ltab(1) := 'TAB1';
        ltab(2) := 'TAB2';
        FOR i IN ltab.first .. ltab.last
        LOOP
            dbms_stats.gather_table_stats(ownname => USER, tabname => ltab(i) , estimate_percent => dbms_stats.auto_sample_size,
            method_opt => 'for all indexed columns size auto', degree =>
            dbms_stats.auto_degree ,CASCADE => TRUE );
        END LOOP;
    END p_manual_gather_table_stats;
    --Scheduled Job
    BEGIN
        -- Job defined entirely by the CREATE JOB procedure.
        DBMS_SCHEDULER.create_job ( job_name => 'MANUAL_GATHER_TABLE_STATS',
        job_type => 'PLSQL_BLOCK',
        job_action => 'BEGIN p_manual_gather_table_stats; END;',
        start_date => SYSTIMESTAMP,
        repeat_interval => 'FREQ=DAILY; BYHOUR=12;BYMINUTE=45;BYSECOND=0',
        end_date => NULL,
        enabled => TRUE,
        comments => 'Job to manually gather stats for tables: TAB1,TAB2. Runs at 12:45 Daily.');
    END;Thanks,
    Somiya

    The question was, is there a better way, and you partly answered it.
    Somiya, you have to be sure the queries have appropriate statistics when the queries are being run. In addition, if the queries are being run while data is being loaded, that is going to slow things down regardless, for several possible reasons, such as resource contention, inappropriate statistics, and having to maintain a read consistent view for each query.
    The default collection job decides for each table based on changes it perceives in the data. You probably don't want the default collection job to deal with those tables. You probably do want to do what Dan suggested with the statistics. But it's hard to tell from your description. Is the data volume and distribution volatile? You surely want representative statistics available when each query is started. You may want to use all the plan stability features available to tell the optimizer to do the right thing (see for example http://jonathanlewis.wordpress.com/2011/01/12/fake-baselines/ ). You may want to just give up and use dynamic sampling, I don't know, entire books, blogs and papers have been written on the subject. It's sufficiently advanced technology to appear as magic.

  • Business Objects Scheduler Jobs Monitoring

    Hi,
    Is there a management pack available for integration with System Centre (SCOM) such that we can monitor failed scheduled jobs in Business Objects via SCOM?
    Is there any other way to monitor failed scheduled Business Objects jobs straight from the database or via any other method?
    Thanks,
    Chandan

    Hi Chandran,
    I'm not a SCOM expert but I understand it is basically a monitoring tool.
    Normally you would check for failed schedules in Business Objects using the Instance Manager, to filter for failed scheduled reports. Alternatively you could use the Query Builder to report directly against the XI3.1 Repository and return a list of failed instances.
    Above workflows can also be coded into the SDK to provide customised feedback on schedule failures
    The final option would be to simply set up email Notifications on the schedules to simply email the Administrator in the event of a schedule failure.
    With any of the above options it should be possible to link in your SCOM system to read\monitor from this output.
    I hope this helps.
    Kind regards,
    John

  • Change print recipient in scheduled jobs (SM37)

    Is it possible to change the print recipient in multiple scheduled jobs (SM37) at the same time?
    Thanks,
    Mark

    You have to change it manually one by one... it is not possible to change it in a multiple jobs at a same time.

  • Analytical snapshots and scheduled jobs

    Hello,
    for extended analytical reports we need to make the snapshots of our statistic datas every day.
    There is an option of doing snapshots of the Project in Project Management, but we need to have snapshots of our own BO's or data sources. Is there any possibility to do this?
    In case snapshots can't be done by the system we also would perform it in coding, by creating business objects with required datas and saving them. But we need to do the saving automatically (regularly).
    In Project Management it is also possible to run scheduled snapshots. How can we schedule jobs (it would be enough if we could execute an action which saves our datas at scheduled time)?
    We have found only "Mass Data Run Process" in this context but the MDR's can be done only for the special standard Floorplans, aren't they?
    Best regards,
    Leonid Granatstein.

    Hi,
    currently it is not possible to create snapshots for partner BO content by a standard process.
    Also defining a mass data run object, which you could schedule and where you could implement the snapshot by yourself, is not available for partners yet.  So we have no standard mechanism available to get your task done with the use of the current implementation possibilities.
    The only option I see is that you develop the snapshot activity on your own in ByD studio.
    To trigger the snapshot activity on a regular basis, I only see the option to trigger this from outside. An option would be to define a web service or an XML file upload.
    You could write for example a small program ( by using PHP or .NET) on a PC which runs on a regular basis and which uploads an XML file or calls a web service. This then triggers the snapshot activity you have programmed in ByD Studio.
    I hope this helps.
    Regards,
    Thomas

  • Report Queue Manager  - Scheduled Jobs

    Question: How/where does the Queue Manager store info for scheduled jobs?

    It is possible (IN 9iAS) to configure the Reports Server to store all requests in a database table. The Table(s) / View(s) are created by the script rw_server.sql located in $ORACLE_HOME/reports/admin/sql
    To configure this options, create the table and view with the rw_server.sql script, then add these lines in the
    <reports server name>.conf file <jobStatusRepository class="oracle.reports.server.JobRepositoryDB"> <property name="repositoryConn" value="user/password@tnsalias" confidential="yes" encrypted="no"/> </jobStatusRepository> The "value" attribute of the property "repositoryConn" should be a user/password@tnsalias that will be used to connect to the database. The account must have access to the tables created by the script rw_server.sql The combination confidential="yes" and encrypted="no" will provoke the encryption of the value attribute once the Report Server is re-started : After the Reports Server has started you will find: <jobStatusRepository class="oracle.reports.server.JobRepositoryDB"> <property name="repositoryConn" value="ihjFFBFD3jk5Xfrd5HeX5w3dE/E=" confidential="yes" encrypted="yes"/> </jobStatusRepository> Description of the rw_server_queue view columns can be found in the script rw_server.sql itself.
    Thanks,

  • Capture DDL for Oracle Scheduled Jobs and job from DBMA_JOBS

    I have trying at this for while and now I am stuck...so I need some help or advice.
    I am doing a DB migration and I need to capture the DDL for all Oracle Scheduled Jobs as well as DBMS_JOBS. Anyone have a good way?
    I took a datapump export last week. Can I run a impdp and capture the DDL for this VIA the sqlfile parameter? If so, how? What would be the script for it. Something like this maybe (par file)?
    directory=DP_IMPORT
    DUMPFILE=dump1.dmp
    SCHEMAS=A,B,C...X,Y,Z
    LOGFILE=JOBS.LOG
    EXCLUDE=TABLE,INDEX,CONSTRAINT
    PARALLEL=4
    INCLUDE=JOBS
    ????

    Hello,
    It's JOB, see following table for possible options , I recommend you to read throught this thoroughly before starting.
    http://download.oracle.com/docs/cd/B19306_01/appdev.102/b14258/d_metada.htm#BGBIEDIA
    Regards

  • Where is the scheduled job?

    This scheduled job starts at 22:00 every day on the Linux server.  it creates snapshot on a remote storage via ssh,  and here is what I found in the log about the job on the remote node:
    Sun Nov 10 22:00:08 EST [remote_node: rshd_0:debug]: [email protected](linux server ip)[email protected](ip)_37397:IN:ssh2_37397:IN:ssh2 shell:SSH INPUT COMMAND is snap create hourly_11-10-2013_22-00
    It seems root cron job is the only cron job under /var/spool/cron on this Linux server. However, I can not find the script which started at 22:00 at all, nor any script with such function. the following is the only line at 22:00pm in /var/cron/log
    Nov 10 22:01:01 phoenix crond[24352]: (root) CMD (run-parts /etc/cron.hourly)
    However, there is no any scripts with such function in /etc/cron.hourly neither.
    Please help me out where this schedule job possibly located on the linux server.
    Thank you in advance!

    Understand all you are saying here.
    I have already talk to NetApp support. They are right that the action on the storage is definitely triggered by this linux server "phoenix". The following message has clearly stated that:
    Sun Nov 10 22:00:25 EST [netappname: rshd_4:debug]: [email protected]_37412:IN:ssh2 shell:SSH INPUT COMMAND is snap create vol1 vol1_nightly_11-10-2013_22-00
    159.3.99.146 is the ip for the linux server "phoenix".
    However, they can not tell me more since it should be found out on the linux.
    The problem is that I could not locate the script either by searching crontab jobs or any anacron as you explained.
    root is only user who has cron job. I checked everysingle line, not only there are no such scripts, also there are no any jobs will ge kicked at 22:00 as the time when the action is started.
    I also searched /etc/cron.hourly. there is only one script, and it does nothing about that action.
    so, i don't have a clue about what and where the script possible is...

  • Empty "scheduled jobs" list in control center

    Hi all,
    due to a move to a new data warehouse environment i added a new configuration
    to my OWB project with new locations, control center etc.
    I was able to deploy and run all of the owb stuff into the new environment except all
    the scheduled jobs to deploy into OEM,
    which owb generates for all the process flows that have a schedule assigned ("Referred Calendar" in Configuration).
    The list of "Scheduled Jobs" in control center for the target location just appears empty.
    Are there any additional settings or privileges missings,
    i.e. for target location, owb repository owner, configuration, location, ...?
    Old env is Oracle DB 10.2.0.2 on Win Server 2003 32 bit, Workflow Server 2.6.4, OWB Client 10.2.0.1.31, OWB-Repository 10.2.0.1.0.
    New env is Oracle DB 10.2.0.2 on Redhat 64 bit, Workflow Server 2.6.4, OWB Client 10.2.0.1.31, OWB-Repository 10.2.0.1.0.
    Any hints would be appreciated.
    Thanks,
    Matthias

    Hi,
    I found out that I have to create a new schedule module.
    In the create module wizard you'll have to define a location where schedules
    are deployed to and there I defined the new target location.
    It is not possible to assign different deployment locations for same schedule
    module in each configuration.
    I don't know if it is a bug or a feature, but with the new seperate schedule module
    it works.
    Just to let you know,
    Matthias

Maybe you are looking for

  • Total frustration organizing photos

    I am a new Mac owner and have to say that, so far, I think iPhoto is the WORST PROGRAM EVER!!!  I have almost 5,000 photos that I transferred over from my PC.  They were all neatly arranged in folders that I set up up according to *MY* preferences...

  • Rounding to Quarter hour

    Newcomer to numbers.. I can't seem to get my roundup logic working. To break this into steps, it's a time analysis, but I seem to be missing something c2=10:00 c3=10:20 goal is to have the equation return .5 here I'd like to break down the logic I'm

  • How to call amethod from some other application

    Hi, I have two web application running on the tomcat say [a] and . I need to call a method from [a] in [b] without adding the reference jar file of [a] in [b]. How to do this? Please guide me.. Thanks in advance.

  • JQuery Cycle issue with IE9

    I'm in the process of reworking an old website and wanting to insert a JQuery Cycle slide show from the CS6 DW widget browser.  The slide show looks beautiful on all browsers except, you guessed it, Internet Explorer. Here's the deal, when loaded on

  • How to completely install the FRC.

    NI engineer:         Hello, it's my honor to have a chance to ask you some questiones. Recently I have been interested in using kinect to do gesture tracking. And I have got a API which is pasted in attachments. And I found that if I want to use this