Report job is not run at scheduled time

OS:WindowsServer2008R2
Oracle BI 11g(11.1.1.6.0)
CT defined a report job
====
Frequency Daily
Start 2013/01/14 AM 06:00:00
End 2020/01/14 PM 11:59:00
timezone [GMT+09:00]Osaka,Tokyo
====
but he found the report is not run at 06:00:00
at 2013/01/16 08:50AM,he login to BIP,and got to Report Job history.
he found the staus of this report is running and
start processing is 2013/01/16 08:51:32AM
It seems the report job is run after login.
In CT's environment , the followinig work are done Every Day
22:00 stop OracleBI  (opmnctl.bat→stopManagedWebLogic.cmd→stopWebLogic.cmd)
00:00 restart server
05:45 clear cashe (nQCmd.exe)
05:45 start OracleBI(BI_START.bat)
06:00 report job is not run for no mail received.-->★
Could you tell me why the report job are not run at 06:00AM
Does any one can give me some advice on investigating this issue, which information are needed?

Any suggestion on this issue?

Similar Messages

  • Scheduled Brio/IR Jobs are not running as expected in EPM 11.1.2.2

    Hi All,
    We are observing some scheduled jobs are not running as expected,they are not running on the scheduled time.
    Given below are more details.
    We have 2 Jobs running at 6:00 AM daily based on single Recurring Time Event and it is observed that this jobs are not running on time(ex: 6:06 AM EST).
    I tried verifying Consolidate Job Status in workspace admin options and the v8_RecurtimeEvent table for any miss in the 'Next Run Time',
    but they were having the scheduled date/time of run.No clue from EPM Logs.
    Interesting observation is the same jobs are running after updating the schedule time a little bit further(ex: 6:06 AM to 6:07 AM),but again after some days
    the updated job(6:07 AM ) will stop running.
    Thanks,

    Put the IR job logs in TRACE:32 mode which will give more information in the logs . Refer this document link to put in TRACE mode :http://docs.oracle.com/cd/E17236_01/epm.1112/epm_install_troubleshooting_1112200.pdf
    Thanks,
    KK

  • Scheduled jobs are not running DPM 2012 R2

    Hi,
    Recently upgraded my dpm 2012 sp1 to 2012 R2 and upgrade went well but i got 'Connection to the DPM service has been lost.(event id:917 and other event ids in the eventlog errors ike '999,997)'. Few dpm backups are success and most of the dpm backups consistenancy
    checks are failed.
    After investigating the log files and found two SQL server services running in the dpm 2012 r2 server those are 'sql server 2010 & sql server 2012 'service. Then i stopped sql 2010 server service and started only sql server 2012 service using (.\MICROSOFT$DPM$Acct).
    Now 'dpm console issue has gone (event id:917) but new issue ocurred 'all the scheduled job are not running' but manully i can able to run all backup without any issues. i am getting below mentioned event log errors 
    Log Name:      Application
    Source:        SQLAgent$MSDPM2012
    Date:          7/20/2014 4:00:01 AM
    Event ID:      208
    Task Category: Job Engine
    Level:         Warning
    Keywords:      Classic
    User:          N/A
    Computer:      
    Description:
    SQL Server Scheduled Job '7531f5a5-96a9-4f75-97fe-4008ad3c70a8' (0xD873C2CCAF984A4BB6C18484169007A6) - Status: Failed - Invoked on: 2014-07-20 04:00:00 - Message: The job failed.  The Job was invoked by Schedule 443 (Schedule 1).  The last step to
    run was step 1 (Default JobStep).
     Description:
    Fault bucket , type 0
    Event Name: DPMException
    Response: Not available
    Cab Id: 0
    Problem signature:
    P1: TriggerJob
    P2: 4.2.1205.0
    P3: TriggerJob.exe
    P4: 4.2.1205.0
    P5: System.UnauthorizedAccessException
    P6: System.Runtime.InteropServices.Marshal.ThrowExceptionForHRInternal
    P7: 33431035
    P8: 
    P9: 
    P10: 
    Log Name:      Application
    Source:        MSDPM
    Date:          7/20/2014 4:00:01 AM
    Event ID:      976
    Task Category: None
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:      
    Description:
    The description for Event ID 976 from source MSDPM cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event: 
    The DPM job failed because it could not contact the DPM engine.
    Problem Details:
    <JobTriggerFailed><__System><ID>9</ID><Seq>0</Seq><TimeCreated>7/20/2014 8:00:01 AM</TimeCreated><Source>TriggerJob.cs</Source><Line>76</Line><HasError>True</HasError></__System><Tags><JobSchedule
    /></Tags></JobTriggerFailed>
    the message resource is present but the message is not found in the string/message table
    plz help me to resolve this error.
    jacob

    Hi,
    i would try to reinstall DPM
    Backup DB
    uninstall DPM
    Install DPM same Version like before
    restore DPM DB
    run dpmsync.exe -sync
    finished
    Seidl Michael | http://www.techguy.at |
    twitter.com/techguyat | facebook.com/techguyat

  • Scheduled jobs do not run as expected after upgrading to 10.2.0.3 or 4

    FYI, we've had a ticket open for several days because our scheduled jobs (dbms_scheduler) would no longer run as scheduled after an upgrade to 10.2.0.4 on HPUX, couldn't find the solution by searching in metalink, nor did I find it here or using Google - obviously I wasn't searching correctly. There is a note id that references a set of steps that appears to have resolved our problem. I am putting this note out here so that if you encountered the same difficulty you may come across this note earlier in your troubleshooting efforts rather than later. The full title of the note is: 'Scheduled jobs do not run as expected after upgrading to either 10.2.0.3 or 10.2.0.4'. The Doc ID is: 731678.1.

    Thanks - our ticket should be getting closed out on it (our dba will be updating it), the scheduler has been running reliably since we took the steps in the doc mentioned.

  • Job not running on expected time

    Hi,
    Please see the job details
    X NUMBER;
    BEGIN
    SYS.DBMS_JOB.SUBMIT
    *( job => X*
    *,what => 'BEGIN INVS_RIM_DATA_LOAD(''01A'',''RIMADMIN'',TRUNC(SYSDATE)); END;'*
    *,next_date => to_date('15/08/2011 23:09:00','dd/mm/yyyy hh24:mi:ss')*
    *,interval => 'TRUNC(SYSDATE+1)+23.15/24'*
    *,no_parse => TRUE*
    SYS.DBMS_OUTPUT.PUT_LINE('Job Number is: ' || to_char(x));
    END;
    commit;
    For taking the cold backup, db is shut down on every Friday 11:00PM and open by 02:30 PM Saturday. So job expected to run on Friday night will be running by Saturday 02:30AM. It is ok for us. But problem is job is not running on Saturday night.
    When checked on Sunday morning the user_jobs shows details as
    Last_date : 13-Aug-2011 02:30:14 AM
    Next_Date : 14-Aug-2011 11:09:00 PM
    Issue we are facing is job is not running on 13-Aug 11:09:00 PM
    Please help me to fix this issue.

    Well, it runs on the expected time (for me). This is your interval: TRUNC(SYSDATE+1)+23.15/24
    If the last job run at 13-Aug-2011 02:30:14 AM, then the next one should be:
    = TRUNC( '13-Aug-2011 02:30:14 AM' + 1 ) + 23.15/24
    = TRUNC( '14-Aug-2011 02:30:14 AM' ) + 23.15/24
    = '14-Aug-2011' + 23.15/24
    = '14-Aug-2011 11:09:00 PM'
    You can create an additional job which will work at Saturday nights. This is the easiest fix I can think.
    Regards
    Gokhan Atil

  • The SSP Timer Job Distribution List Import Job was not run.

    The SSP Timer Job Distribution List Import Job was not run.   Reason: Logon failure: the user has not been granted the requested logon type at this computer
    We are facing the above mentioned problem . I have added the service account in backup operates and did everything like update farm credentils , update managed account etc , but still no luck . Please note that we haven't done any chnages to the account
    Application pool is keeping going into stop mode when we try to access the site . Both CA and webapplication are down with the error as Service Unavailble . Please help
    Srini

    sounds like the account can only be used on certain COMPUTERS ("has not been granted the requested logon type
    at this computer")... check active directory.
    Scott Brickey
    MCTS, MCPD, MCITP
    www.sbrickey.com
    Strategic Data Systems - for all your SharePoint needs

  • IDOC Monitoring issue - job BPM_DATA_COLLECTION* not running.

    Hi all,
    We are facing an issue with BPM "IDOC Monitoring" (under application monitoring), which we have setup to monitor Inbound and Outbound Idocs in 2 separate R/3 systems.
    In one system it works fine, and measured values are returned each time the monitor is set to run according to the specified schedule.
    However, for another R/3 system, the monitor has never run, even though the settings are identical in the monitoring setup.
    From reading the Interfaces Monitoring Setup guide, I found that this monitor depends on a job called BPM_DATA_COLLECTION* which runs in the monitored system. In the system where monitoring is functioning correctly, I can see this job is completing successfully at the time the monitor is set to run. All I find is completed jobs - no scheduled or released jobs present.
    However, for the system where the monitoring is not functioning, I found that this job is not running, but that the job is sitting in scheduled status instead.
    When I tested manually running the job in this system, it ran successfully, and in Solution Manager the monitor brought back the measured value, so the monitor only ran successfully when I manually ran the job in the monitored system.
    I read notes 1321015 & 1339657 relating to IDOC monitoring. 1321015 appears to be more relevant, yet it does not exactly describe my issue - it mentions the job BPM_DATA_COLLECTION* failing rather than just remaining in scheduled status which is what I see.
    Anyone else see this issue before?
    On a more general point - the standard BPM Setup guide doesn't really go into much detail on IDOC Monitoring, and makes no mention of what is happening in the background, i.e. the job BPM_DATA_COLLECTION* being created and run as per schedule. This info is found in a separate document "Interface Monitoring Setup Guide".
    Is there any single document which describes fully what happens both in the Solution Manager and the Monitored systems when BPM is activated? For example, to describe which monitors require jobs to be run, which monitors require additional setup in monitored system, etc? A document such as this which describes exactly the process flow for each monitor would be very useful in troubleshooting issues going forward.
    Thanks,
    John

    Hello John,
    most probably the user assigned to the corresponding RFC READ connection that connects SolMan with the backend system doesn't have proper authorization to release a job. That's why it is only created/scheduled but not released. Verify if the RFC user on the backend has the latest CSMREG profile assigned according to SAP note 455356.
    You can also check if the latest ST-PI support package is installed  on your backend system as the ST-PI usualy contain the latest definition of CSMREG.
    Best Regards
    Volker

  • Background Job is not running in KW

    Hi All,
    I have problem with background job.
    I am working on Knowledge warehouse server.
    And I am scheduling background job to upload contents.
    But background job is not finishing at proper time.
    And while displaying trace it is giving following error.
    ERROR => BtcCleanUp: BtcLgAp-call failed (rc = 4) [btcjcntl.c   1251]
    Please help me to resolve this.
    Regards,
    Payal patel

    Hi Juan,
    I have tried with process as well as instance restart.
    It didn't help me.
    Hi Santosh,
    Following is the part of trace file (level -2).
    First line shows error.
    Tell me if you need any other log file.
    L  *** ERROR => BtcCleanUp: BtcLgAp-call failed (rc = 4) [btcjcntl.c   1251]
    M  read msgserver-list from MBUF
    M  ThSemRq (4, 1, 0, 0)
    M  ThSemRel (4, 1)
    M  ThSetBtcName: found batch server sapbl4_KW7_00                          
    M  ThScheduler2: server name: sapbl4_KW7_00                          
    M  ThISendMsg: send message (5) to server (wp) with name >sapbl4_KW7_00                           <
    M  ThISndName: send to name: >sapbl4_KW7_00                           <
    M  ThISend: (tm/type/info = 10/0x2000/0x0, mode_deleted=0)
    M  ThRqOutCheck: o.k.
    M  abap strategy ROLL / O.K.
    M  ThNewWpStat (type=0x2000, task_switch=0, inline_hold=0, hand_shake=0, debug=0, ..)
    M  ThNewWpStat: new state of T10/M0 = 0x3c
    M  ThISend: new wp stat: 0x0
    M  Adresse   Offset  Message by name (one way)
    M  -
    M  06C60600  000000  00000000 05000000 255f4556 454e545f |........%_EVENT_|
    M  06C60610  000016  53434845 44554c45 52202020 20202020 |SCHEDULER       |
    M  -
    M  ThMkReq: send output to canceled mode
    M  make DISP owner of wp_ca_blk 180
    M  DpRqPutIntoQueue: put request into queue (reqtype 0, prio LOW, rq_id 19358)
    M  -OUT- sender_id WORK_PROCESS      tid  10    wp_ca_blk   180     wp_id 10
    M  -OUT- action    SEND_MSG_ONEWAY   uid  11    appc_ca_blk -1      type  NOWP
    M  -OUT- new_stat  NO_CHANGE         mode 0     len         268     rq_id 19358
    M  -OUT- forward   DIA              
    M  -OUT- req_info  CANCELMODE MSG_WITH_REQ_BUF MSG_WITH_OH
    M  nihsl-getHostAddr: got hostname 'localhost' from operation system
    M  nihsi-getHostAddr: hostname 'localhost' = addr 127.0.0.1
    M  nihsl-getServNo: got servicename 'sapdp00' from operation system
    M  nihsi-getServNo: servicename 'sapdp00' = port 0C.80/3200
    M  NiIInitSocket: set default settings for socket 1196
    M  NiCreateHandle: state hdl 1 / socket 1196 NI_INITIAL
    M  NiIDgSend: init datagram send hdl 1 / socket 1196
    M  CPU byte order: little endian, reverse network, low val .. high val
    M  NiIDgSend: connect dgram to: host 127.0.0.1, port 0C.80/3200, fam 2 (low adr..high adr)
    M  NiIDgsend: sending on connected datagram-handle
    M  LOCK WP ca_blk 181
    M  ThResFree: free resources of U11 M0 I2 (normal mode, complete free) at level 3, errno=23, db_action=TH_DB_NO_ACTION, pooling=1
    M  ThResFree: reset spa state for user T10/U11/M0
    M  ThCallHooks: call hook >RtmClearSession< for event BEFORE_SESSION_CANCEL
    M  ThCallHooks: call hook >HttpClearSession< for event BEFORE_SESSION_CANCEL
    M  ThCallHooks: call hook >SpoolHandleHook< for event BEFORE_SESSION_CANCEL
    M  SosSearchAnchor: search anchor for 2
    M  PfStatDisconnect: disconnect statistics
    M  ThDealComm: del 1 cpic conn(s) of T10/U11/M0 (errno/deal_r2/free_level = 23/1/3)
    M  ThCPICFree: send DEAL to U11/M0
    M  ThICMDEAL (14030609, ..)
    M  ThCPIC: execute cpic func DEALLOCATE
    M  ThCPIC: last_ftype/last_timeout/last_requested_length/last_receive_data 1/-1/-1/0
    M  ThCPIC: allowed rq_type of T10/M0 = TH_APPC_RC_RQ
    M  ThConnectToLocGw: connect to local gateway
    M  GwConnectSapWp: connect to gateway >localhost< >sapgw00<
    M  nihsl-getHostAddr: found hostname 'localhost' in cache
    M  nihsi-getHostAddr: hostname 'localhost' = addr 127.0.0.1
    M  nihsl-getServNo: got servicename 'sapgw00' from operation system
    M  nihsi-getServNo: servicename 'sapgw00' = port 0C.E4/3300
    M  NiCreateHandle: state hdl 2 / socket -1 NI_INITIAL
    M  NiIBlockMode: switch off block-mode for hdl 2 / socket -1
    M  NiLowLevCon: connect to: host 127.0.0.1, port 0C.E4/3300, fam 2 (low adr..high adr)
    M  NiIInitSocket: set default settings for socket 1140
    M  NiISocket: hdl 2 got socket 1140
    M  NiPConnect: connect in progress
    M  SiPeekPendConn: connection of socket 1140 established
    M  NiLowLevCon: took local port 0E.51/3665
    M  nilh-localCheck: using local address list
    M  NiSetStat: state hdl 2 NI_CONNECTED
    M  NiIWrite: write 64, 1 packs, MESG_IO, hdl 2, data complete
    M  NiIPeek: peek successful for hdl 2 / socket 1140 (r)
    M  NiIRead: read 64, 2 packs, MESG_IO, hdl 2, data complete
    M  GwConnectSapWp: connect to gateway localhost / sapgw00 (pid = 6184) o.k.
    M  ThISend: (tm/type/info = 10/0x80/0x0, mode_deleted=0)
    M  ThRqOutCheck: o.k.
    M  abap strategy ROLL / O.K.
    M  ThNewWpStat (type=0x80, task_switch=0, inline_hold=0, hand_shake=0, debug=0, ..)
    M  ThNewWpStat: new state of T10/M0 = 0x3c
    M  ThISend: new wp stat: 0x0
    Regards,
    Payal Patel
    Message was edited by:
            Payal Patel

  • Job is not running in Source system.

    Hi Experts,
    One issue I have because of this I am not able to load data into Data sources.
    I am in BI 7.0 environment.
    when I execute infopackage, total and techinical status are in yellow.
    I found in R/3 Job is not running, based on this statement in SM37. from the fallowing.
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 0 records
    Result of customer enhancement: 0 records
    IDOC: Info IDoc 2, IDoc No. 3136, Duration 00:00:00
    IDoc: Start = 14.05.2009 07:23:39, End = 14.05.2009 07:23:39
    Synchronized transmission of info IDoc 3 (0 parallel tasks)
    IDOC: Info IDoc 3, IDoc No. 3137, Duration 00:00:00
    IDoc: Start = 14.05.2009 07:23:39, End = 14.05.2009 07:23:39
    Job finished
    Job satrt time and end time are same, so job is not running in the sourcesystem, am i right. let me known if i am worng.
    It is Standard Data source but in the above statement becasue  *result of Customer enhancement:0 records.*
    please help me to reslove this issue to load data into BI.
    1. what I have to do to run job in source system side.
    2. should I take any help from Basis.
    Regards
    Vijay
    Edited by: vijay anand on May 14, 2009 3:02 PM
    Edited by: vijay anand on May 14, 2009 3:04 PM

    Hi Rupesh,
    in RSA3 data is availbale,
    but it is not coming to BI,
    Messages from source system
    see also Processing Steps Request
    These messages are sent by IDoc from the source system. Both the extractor itself as well as the service API can send messages. When errors occur, several messages are usually sent together.
    From the source system, there are several types of messages that can be differentiated by the so-called Info-IDoc-Status. The IDoc with status 2 plays a particular role here; it describes the number of records that have been extracted in a source system and sent to BI. The number of the records received in BI is checked against this information.
    the abovemessage I am getting in details tab.
    Regards
    Vijay

  • Oracle automatic statistics optimizer job is not running after full import

    Hi All,
    I did a full import in our QA database, import was successful, however GATHER_STATS_JOB is not running after sep 18 2010 though its enable and scheduled, i did query last_analyzed table to check and its confirmed that it didnt ran after sep18,2010.
    Please refer below for the output
    OWNER JOB_NAME ENABL STATE START_DATE END_DATE LAST_START_DATE NEXT_RUN_D
    SYS GATHER_STATS_JOB TRUE SCHEDULED 18-09-2010 06:00:02
    Oracle defined automatic optimizer statistics collection job
    =======
    SQL> select OWNER,JOB_NAME,STATUS,REQ_START_DATE,
    to_char(ACTUAL_START_DATE, 'dd-mm-yyyy HH24:MI:SS') ACTUAL_START_DATE,RUN_DURATION
    from dba_scheduler_job_run_details where
    job_name='GATHER_STATS_JOB' order by ACTUAL_START_DATE asc; 2 3 4
    OWNER JOB_NAME STATUS REQ_START_DATE ACTUAL_START_DATE
    RUN_DURATION
    SYS GATHER_STATS_JOB SUCCEEDED 16-09-2010 22:00:00
    +000 00:00:22
    SYS GATHER_STATS_JOB SUCCEEDED 17-09-2010 22:00:02
    +000 00:00:18
    SYS GATHER_STATS_JOB SUCCEEDED 18-09-2010 06:00:02
    +000 00:00:26
    What could be the reason for GATHER_STATS_JOB job not running although its set to auto
    SQL> select dbms_stats.get_param('AUTOSTATS_TARGET') from dual;
    DBMS_STATS.GET_PARAM('AUTOSTATS_TARGET')
    AUTO
    Does anybody has this kind of experience, please share
    Apprecitate your responses
    Regards
    srh

    ?So basically you are saying is if none of the tables are changed then GATHER_STATS_JOB will not run, but i see tables are updated still the job is not running. I did >query dba_scheduler_jobs and the state of the job is true and scheduled. Please see my previous post on the output
    Am i missing anything here, do i look for some parameters settings
    So basically you are saying is if none of the tables are changed then GATHER_STATS_JOB will not run,GATHER_STATS_JOB will run and if there are any table in which there's a 10 percent change in data, it will gather statistics on that table. If no table data have changes less than 10 percent, it will not gather statistics.
    http://download.oracle.com/docs/cd/B19306_01/server.102/b14211/stats.htm#i41282
    Hope this helps.
    -Anantha

  • Rule set import - Background job did not run

    Hi,
    I am setting up my CC 5.2 production system. I have downloaded from ruleset from the dev CC and imported it into production. However the background job generated did not run. I am implementing SAP note 999785 to fix this, but am wondering what should I do about the rule set? Do I need to delete the rule set and reimport it? As this background job did not run I notice that the permission rules did not generate.
    Any advice is welcome.
    Thanks

    Hi,
    as the backgorund job never ran and no rules were created, I can just reimport the ruleset and let the job run. I have tried this and the rules were created successfully

  • SQL Server Agent Job is not running as per schedule

    I have a job that is supposed to run every 6 AM.  It was running last week but not this morning. 
    The big problem is that there are no error messages what so ever. I checked Agent History, SQL server logs and Windows system logs.
    I checked the schedule and it is set to run everyday and there are no limits on how long it is supposed to run and it is enabled.
    Did anyone get this problem before?

    Hi,
    Sometimes there are valid reasons a job will NOT kick-in on the scheduled time. Sometime ago, our database full backup job didn't start on time and in fact just
    got skipped over. Upon further investigation, a idiot started doing defragging on the target (destination) server without informing anybody in the enterprise and the previous days backup was running over 24+ hours instead of 2 hours. Sometime people cheat
    and try to enable/disable jobs and in those cases you may want to add notification if something like this happens. http://www.mssqltips.com/sqlservertip/1803/auto-alert-for-sql-agent-jobs-when-they-are-enabled-or-disabled/
    Sankar Reddy
    Blog: http://SankarReddy.com/
    Twitter: http://twitter.com/SankarReddy13/

  • How to Schedule Jobs to only run during a time window

    I have a long running task that needs to schedule jobs to process data.
    I only want these scheduled jobs to start during a specific window of time each day, probably 10:00 PM to 6:00 AM.
    If the scheduled jobs do not begin during the specified time frame, they must wait until the next day to start running.
    Each scheduled job will only be executed once and then auto dropped.
    How should I go about creating these scheduled jobs?

    Hi Jeff,
    I agree that the documentation isn't clear enough about the purpose of windows.
    You can indeed use windows for changing the resource plan, but you can also use them for scheduling your jobs.
    I did a simple test in real-time to illustrate the latter.
    At around 10.30 am today I created a table that will populated by a job:
    CREATE TABLE TEST_WINDOW_TABLE(EVENT_DATE DATE);
    Then, I created a window whose start_date is today at 10.40 am :
    dbms_scheduler.create_window(
                                 window_name     =>'TEST_WINDOW',
                                 resource_plan   => NULL,
                                 start_date      => to_date('10/04/2014 10:40:00', 'dd/mm/yyyy hh24:mi:ss'),
                                 repeat_interval => NULL,
                                 duration        =>interval '5' minute
    You can see that this window doesn't have a resource plan, and its repeat interval is NULL (so it will be opened only once).
    The window will stay open for 5 minutes.
    Finally, I created a one-off job whose schedule is the previously created window:
    DBMS_SCHEDULER.create_job (
                               job_name      => 'TEST_WINDOW_JOB',
                               job_type      => 'PLSQL_BLOCK',
                               job_action    => 'BEGIN insert into test_window_table values (sysdate); COMMIT; END;',
                               schedule_name => 'SYS.TEST_WINDOW',
                               enabled       => true,
                               auto_drop     => true
    Checking the user_scheduler_job_log before 10.40 would return no rows, which mean the job hasn't started yet since the window was not open.
    Now, from 10.40, it shows one entry:
    SQL> select log_date, status from user_scheduler_job_log where job_name = 'TEST_WINDOW_JOB';
    LOG_DATE                                                                         STATUS
    10/04/14 10:40:02,106000 +02:00                                                  SUCCEEDED
    The TEST_WINDOW_TABLE has also got the row:
    SQL> select * from TEST_WINDOW_TABLE;
    EVENT_DATE
    10/04/2014 10:40:02
    Voilà.
    In your case, since you want to run the jobs daily between 10 pm and 6 am (duration of 8 hours), the window would look like this:
    dbms_scheduler.create_window(
                                 window_name     =>'YOUR_WINDOW',
                                 resource_plan   => NULL,
                                 repeat_interval => 'freq=daily;byhour=22;byminute=0;bysecond=0',
                                 duration        =>interval '8' hour
    For your jobs, you may need to specify an end_date if you want to make sure the job gets dropped if it couldn't run in its window.

  • Svrcfg show Job Server not running

    Hi,
    I am currently running Data Services 12.1.1.0 on Linux RH AS 4.6. I have noticed an issue and wondering whether there is something I am doing wrong.
    I installed DS and configured the JS. I have successfully started, stopped and started this again, via the svrcfg utility and it clearly states "running". Then I log out of Putty and log back into the into server hosting DS. When I go to the svrcfg utility again, go to control job server it reports a status of not running. If I exit the tool and type ps -fu <username> I can see my Job Servive/Engine as I would expect and jobs still run fine?
    Does anyone know what might be causing this?
    Thanks
    Chris

    Hi,
    I am also getting the same issue. If you login into putty and check the status by which terminal you have started shows "RUNNING", if you login through another session with same user it shows "NOT RUNNING".
    Output of ps -o user,fname |grep -v grep|grep AL_JobSe is
    bodi     AL_JobSe
    I am running the . .al_env.sh manually (not from profile!).
    Thanks
    Maran

  • Dbms_schduler job is not running on a 2 node rac when 1st node fails

    Hi,
    I want to create a dbms_scheduler job in a 2 node RAC and the job should always run on the node1 and if node1 is down then it should run on node2. This is Oracle 10gR2 (10.2.0.3 in WINDOWS) .In order to do the same I did following
    -- First Step
    Using DBCA- Service Managment - Created a service (BATCH_SERVICE) and given node1 as preferred and node2 as available. This created following entry in tnsnames.ora in both nodes.
    BATCH_SERVICE =
    (DESCRIPTION =
    (ADDRESS = (PROTOCOL = TCP)(HOST = node1-vip)(PORT = 1521))
    (ADDRESS = (PROTOCOL = TCP)(HOST = node2-vip)(PORT = 1521))
    (LOAD_BALANCE = yes)
    (CONNECT_DATA =
    (SERVER = DEDICATED)
    (SERVICE_NAME = BATCH_SERVICE)
    (FAILOVER_MODE =
    (TYPE = SELECT)
    (METHOD = BASIC)
    (RETRIES = 180)
    (DELAY = 5)
    --- Step 2
    -- Created BATCH job classes.
    BEGIN
    DBMS_SCHEDULER.create_job_class(
    job_class_name => 'BATCH_JOB_CLASS',
    service => 'BATCH_SERVICE');
    END;
    -- Step 3 -- created a job using job_class as BATCH_JOB_CLASS
    begin
    dbms_scheduler.create_job(
    job_name => 'oltp_job_test'
    ,job_type => 'STORED_PROCEDURE'
    ,job_action => 'schema1.P1'
    ,start_date => systimestamp at time zone 'US/Central'
    ,repeat_interval => 'FREQ=DAILY;BYHOUR=11;BYMINUTE=30;'
    ,job_class => 'BATCH_JOB_CLASS'
    ,enabled => TRUE
    ,comments => 'New Job.');
    end;
    Now when I monitor this job it runs on node1. Now I started testing for failover. I manually shutdown 1st instance. Then as per my understanding job should run on 2nd node. But job is not picking up.
    when I run the followign command
    srvctl status service -d db -s BATCH_SERVICE
    service BATCH_SERVICE is running on instance node2.
    Any help is really appreciated.

    It does not show that whether job is running or broken.

Maybe you are looking for

  • Manual override option for Unit selling price in Order line level.

    In OM R-12 Our client wants manual override option for Unit selling price in Order line level. But there is no solution available to fulfill this requirment. To be able to manually override the price on a Sales Order line you must have defined a manu

  • Suppress AWT

    I have extended a awt class in my java application. Somehow, it pop up a GUI window. Is there any parameter I can pass to the JVM so it will ignore all the GUI bits? I just want to have a pure console java application. Many Thanks! Art.

  • Alter database recover to logical standby is hanging

    Hi, when I try to convert my physical standby to logical standby using alter database recover to logical standby STANDBYL; the command just hangs I have seen the usual hit on this which is to create the Redo Dictionary on the Primary using execute db

  • Controlling extended classic

    We are using the Extended classic scenario in SRM 5. A requirement has now surfaced where all catalogue items should be processed within SRM PO wise so when the cart is approved it reacts just like extended classic where a copy of the PO goes to R/3.

  • Can't read a delimiter

    Guys, I am reading in a text file, and I am trying to write an interface to read in the different words in there...so I am using a tokenizer for the delimiter...in the text file the delimiter shows as |, but when the file is read in java the string l