WAAS not running preposition jobs

I have a NME-WAE-502-K9 in a 2821 router and have it connected via a ipsec VPN tunnel through a ASA5510
Path to client
CM - router(3925) - switch (4501 with VLAN) - ASA - router 2821 - wae mondule - user
I can ping from the central manager to the WAE unit
I am getting TFO and DR compression but can not get the unit to complete any preposition jobs.
The will show "in progress 0 bytes" after even 24 hours.
IP insepect WAAS enabled is on the router and the ASA 5510
MTU is set to 1500
Totaly at a loss as to were to start looking.
Thanks

When you do prepositioning the Branch WAE acts as the client and goes direct to the server.   As such the Branch WAE IP must be allowed through your Firewall, AND it must get redirected to the Core WAAS accelerator.   Thus, if you are using WCCP to redirect to the Core WAAS device and if you're using a redirect list, make sure to permit the IP of the WAE.
Also you'll want to make sure the server you are trying to preposition from does not have any security restrictions that would prevent the WAE from browsing the directories / copying the files.
Hints to the problem could be found in the errorlog/cifsao-errorlog.current file.
I'd suggest opening a TAC case if the above does not help.

Similar Messages

  • Azure Compute Nodes not running the job

    I have an on-premise head node. I have joined 10 Azure Compute Nodes via the cloud service and storage account. I have uploaded the dlls directory to storage account and synced using hpcsync all the compute nodes. The Azure compute nodes are still not
    running the job. I see in the HPC Job Manager that the Cores In Use = 0. How should I resolve this issue?

    Hello,
    We are researching on the query and would get back to you soon on this.
    I apologize for the inconvenience and appreciate your time and patience in this matter.
    Regards,
    Azam khan

  • Server not running SSIS job in 32bit mode

    I have a package that I developed in VS2012 (using the SQL Data Tools component) which collects data from a DBF file using the VFPOLEDB provider, and puts it into a database on a SQL Server 2012 X64 server. The project containing the package has the DebugOption
    of Run64BitRuntime set to false. I have imported this package to the SSIS Package Store of a test and live server (identical setups). The VFPOLEDB provider is installed on both, and I can see in the registry of both machines that it's there for 32bit runtimes.
    The package runs fine on the test machine, but fails on the live machine. The live instance of SQL does not appear to recognize the 32bit VFPOLEDB provider that is installed.
    The only difference in the instances of SQL is that the live environment has an Integration Service Catalog set up, where as the test does not. Looking at the logs of the servers, when the live starts, it runs sp_ssis_startup, and then logs messages about
    unsafe assemblies being loaded in. This SP is not run on the test environment because there is no catalog.
    The jobs I create have the flag set to use the 32bit runtime, but I can't help but feel that the SSIS Catalog is having an issue with the VFPOLEDB I am using, and not loading it.
    I don't really know anything about the SSIS catalog, so is anyone able to suggest any direction I could move in?

    I am not sure what you want in terms of job step setup information.
    The history is a single message:
    Message
    Executed as user: MDA\catbert. Microsoft (R) SQL Server Execute Package Utility  Version 11.0.5058.0 for 64-bit  Copyright (C) Microsoft Corporation. All rights reserved.    Started:  6:52:01 PM  Error: 2015-02-03 18:52:01.10    
    Code: 0xC0209303     Source: NEF Transfer SSIS2012 Connection manager "Source Connection"     Description: The requested OLE DB provider VFPOLEDB.1 is not registered. If the 64-bit driver is not installed,
    run the package in 32-bit mode. Error code: 0x00000000.  An OLE DB record is available.  Source: "Microsoft OLE DB Service Components"  Hresult: 0x80040154  Description: "Class not registered".  End Error 
    Error: 2015-02-03 18:52:01.10     Code: 0xC020801C     Source: Data Flow Task OLE DB Source [100]     Description: SSIS Error Code DTS_E_CANNOTACQUIRECONNECTIONFROMCONNECTIONMANAGER.  The AcquireConnection
    method call to the connection manager "Source Connection" failed with error code 0xC0209303.  There may be error messages posted before this with more information on why the AcquireConnection method call failed.  End Error  Error:
    2015-02-03 18:52:01.10     Code: 0xC0047017     Source: Data Flow Task SSIS.Pipeline     Description: OLE DB Source failed validation and returned error code 0xC020801C.  End Error  Error:
    2015-02-03 18:52:01.10     Code: 0xC004700C     Source: Data Flow Task SSIS.Pipeline     Description: One or more component failed validation.  End Error  Error: 2015-02-03 18:52:01.10    
    Code: 0xC0024107     Source: Data Flow Task      Description: There were errors during task validation.  End Error  DTExec: The package execution returned DTSER_FAILURE (1).  Started:  6:52:01
    PM  Finished: 6:52:01 PM  Elapsed:  0.094 seconds.  The package execution failed.  The step failed.

  • Use_current_session = FALSE does not run my job correctly

    We have a custom scheduler that invokes jobs based on schedules/conditions. Until now, the jobs were all kicked off in the same session. Since the record set to be processed is increasing, we want the jobs submitted in parallel.
    So the main job is split into several discrete jobs & run them in different sessions (dbms_scheduler.run_job with use_current_session = FALSE). The programs & jobs get created successfully.
    The program has around 12 arguments defined.
    The jobs run; however error out with "ORA-06502: PL/SQL: numeric or value error ORA-06502: PL/SQL: numeric or value error: character to number conversion error" *(DBA_SCHEDULER_JOB_RUN_DETAILS)*
    If I run the jobs with this parameter = TRUE the jobs run successfully. Any pointers greatly appreciated.
    Here are additional details..
    DB: Oracle Database 10g Enterprise Edition Release 10.2.0.1.0 - Prod
    dba_scheduler_global_attribute
    MAX_JOB_SLAVE_PROCESSES
    LOG_HISTORY 30
    DEFAULT_TIMEZONE US/Pacific
    LAST_OBSERVED_EVENT
    EVENT_EXPIRY_TIME
    CURRENT_OPEN_WINDOW WEEKEND_WINDOW
    v$parameter where name like '%process%'
    processes 150
    gcs_server_processes 0
    db_writer_processes 1
    log_archive_max_processes 2
    job_queue_processes 20
    aq_tm_processes 0
    Thanks
    Kiran.

    Hi,
    This error seems clear,
    character to number conversion error : at "XXA.XXX_ANP_ENGINE_MASTER_PKG", line 24
    This is application code which the scheduler did run but the application code is throwing the error.
    You will have to debug the issue occurring at line 24 of package "XXA.XXX_ANP_ENGINE_MASTER_PKG". You may be relying on something in your session which is not available in a background session - so the job fails when run in the background.
    Hope this helps,
    Ravi.

  • Job not running in DBA_Jobs

    Hi,
    We have a job set in dba_jobs.
    Suddenly after 31st mar-2011 the job is not running
    The job of 31st mar 2011 is still in active state and exists in the database
    The details of the job are as follows:
    SELECT r.sid, r.job, r.this_date, r.this_sec, SUBSTR(what,1,40) what
    FROM dba_jobs_running r,dba_jobs j
    WHERE r.job = j.job;
    SID JOB THIS_DATE THIS_SEC
    439 65709 31-MAR-11 06:31:54
    A OS process also exists
    oracle 11130 1 0 Mar 31 ? 1:51 ora_j000_GAT
    SQL> select status, to_char(logon_time,'dd-mm-yyyy:hh24:mi:ss') from v$session where sid=439;
    STATUS TO_CHAR(LOGON_TIME,
    ACTIVE 31-03-2011:06:31:29
    SQL> select THIS_DATE,NEXT_DATE from dba_jobs where job=65709;
    THIS_DATE NEXT_DATE
    31-MAR-11 31-MAR-11
    In long running SQL
    439 SELECT TO_CHAR(HPEOP_EIN) EMPLID, NVL(HP GAT_OWNER 31-03-11:06:31:29 ACTIVE
    EOP_GRADE,' ') JOBCODE, NVL(HPEOP_PREFER
    RED_PHONE,' ') BT_PREFERRED_COMMS, NVL(H
    The jobs usually take 1-5 mins
    Can you please advice what needs to be done
    What could have happened.

    Can you kill the job from the OS?
    Then stop the job from running : set it to broken
    Find out why it doesn't stop this time: look in code and data.
    If you are confident it will not happen again schedule the job again.

  • Cron job not running - script uses rsync, ssh (with id_rsa)

    Does anyone know how to get cron to run a rsync/ssh script which uses id_rsa for passwordless logins?
    I have a backup script I have slaved for months over that works wonderfully when running it from the command line.
    It uses rsync and connects with SSH (passwordless using ssh keys). I have setup id_rsa and id_rsa.pub for passwordless logins (this is my user account, and connects using root@ to the server)
    As stated all works find when run from the command line. But when I setup a cron job to run this file (exact same line I run from the command line) it does not execute. Actually, it does run, as I have an email notifiction which sends to me. It is just the rsync/ssh part that does not run.
    I have spent days and days trying to figure it out. No error messages occur, nothing.  Took me a while to find out cron is emailing the user account on Lion the messages (not to a proper email address).
    The error shows a permission denied, which I believe to be the cron user cannot use the id_rsa/ssh connection.
    Cron is running the script as my user, the same I use on the command line. I even tried copying the id_rsa to the root user .ssh folder in case cron is running as root but that does not work either.
    So does anyone know how to get cron to run this rsync/ssh script which used id_rsa for passwordless logins?
    Thanks a lot

    YES! Got it working, finally!
    So the issue is indeed cron/launchd not being able to access the ssh-agent which holds the key.
    So this is the solution...
    Download keychain for Mac:
    http://www.funtoo.org/wiki/Keychain
    Then once its installed and configured, the script to run needs this added to the top:
    eval `keychain --noask --eval id_rsa` || exit 1
    Note I am using id_rsa whereas keychains instructions refer to id_dsa so I just changed it and it works OK.
    When the script runs the first time it asks for the key password. Then it works from then on.
    What I do not know yet is will I need to re-enter the password when I logoff/logon again. That will cause an issue as I auto restart the Mac every day which means if I am off for a few days backup will stop (can you see the madness in this issue!).

  • Report job is not run at scheduled time

    OS:WindowsServer2008R2
    Oracle BI 11g(11.1.1.6.0)
    CT defined a report job
    ====
    Frequency Daily
    Start 2013/01/14 AM 06:00:00
    End 2020/01/14 PM 11:59:00
    timezone [GMT+09:00]Osaka,Tokyo
    ====
    but he found the report is not run at 06:00:00
    at 2013/01/16 08:50AM,he login to BIP,and got to Report Job history.
    he found the staus of this report is running and
    start processing is 2013/01/16 08:51:32AM
    It seems the report job is run after login.
    In CT's environment , the followinig work are done Every Day
    22:00 stop OracleBI  (opmnctl.bat→stopManagedWebLogic.cmd→stopWebLogic.cmd)
    00:00 restart server
    05:45 clear cashe (nQCmd.exe)
    05:45 start OracleBI(BI_START.bat)
    06:00 report job is not run for no mail received.-->★
    Could you tell me why the report job are not run at 06:00AM
    Does any one can give me some advice on investigating this issue, which information are needed?

    Any suggestion on this issue?

  • Not able to run the job with user id - is

    Hello experts,
    We have problem.
    Every day we run the job.
    ( the job contians two programs called  ZTIBCOPRG  and J_5HJSTP  )).
    I would like to know why the above job is running with the user ID TIBCOADM.  Because this user has German settings for the size conversion in program ZTIBCOPRG and we are having some issues.
    We changed the user to TIBCOUSA and the job would not run. 
    We have changed the user back to TIBCOADM.  It is running.
    KINDLY HELP ME WITH YOUR VALUABLE inputs.   YOUR HELP WILL BE HIGHLY APPRECIATED.
    I have checked for both users in AGR_USERS  table.
    The AGR_USERS is the old user and have many roles compared to userid TIBCOUSA.
    WHERE WE NEED TO SEE . IS it authorisation problem or the other user with whom unable to run the job ?
    or is this the error in the program ZTIBCOPRG which is not able to do size conversion and causing some issues.
    Thanks and Regards,

    Hi,
         after running the job with user id TIBCOUSA, it should be in cancelled state as per your comments.
    you just select the cancelled job and type JDBG in command box and enter it takes you to Debug mode.
    There you may get some information where it is failing.
    Sudheer. A

  • Job is not running in Source system.

    Hi Experts,
    One issue I have because of this I am not able to load data into Data sources.
    I am in BI 7.0 environment.
    when I execute infopackage, total and techinical status are in yellow.
    I found in R/3 Job is not running, based on this statement in SM37. from the fallowing.
    Call customer enhancement EXIT_SAPLRSAP_001 (CMOD) with 0 records
    Result of customer enhancement: 0 records
    IDOC: Info IDoc 2, IDoc No. 3136, Duration 00:00:00
    IDoc: Start = 14.05.2009 07:23:39, End = 14.05.2009 07:23:39
    Synchronized transmission of info IDoc 3 (0 parallel tasks)
    IDOC: Info IDoc 3, IDoc No. 3137, Duration 00:00:00
    IDoc: Start = 14.05.2009 07:23:39, End = 14.05.2009 07:23:39
    Job finished
    Job satrt time and end time are same, so job is not running in the sourcesystem, am i right. let me known if i am worng.
    It is Standard Data source but in the above statement becasue  *result of Customer enhancement:0 records.*
    please help me to reslove this issue to load data into BI.
    1. what I have to do to run job in source system side.
    2. should I take any help from Basis.
    Regards
    Vijay
    Edited by: vijay anand on May 14, 2009 3:02 PM
    Edited by: vijay anand on May 14, 2009 3:04 PM

    Hi Rupesh,
    in RSA3 data is availbale,
    but it is not coming to BI,
    Messages from source system
    see also Processing Steps Request
    These messages are sent by IDoc from the source system. Both the extractor itself as well as the service API can send messages. When errors occur, several messages are usually sent together.
    From the source system, there are several types of messages that can be differentiated by the so-called Info-IDoc-Status. The IDoc with status 2 plays a particular role here; it describes the number of records that have been extracted in a source system and sent to BI. The number of the records received in BI is checked against this information.
    the abovemessage I am getting in details tab.
    Regards
    Vijay

  • Scheduled jobs are not running DPM 2012 R2

    Hi,
    Recently upgraded my dpm 2012 sp1 to 2012 R2 and upgrade went well but i got 'Connection to the DPM service has been lost.(event id:917 and other event ids in the eventlog errors ike '999,997)'. Few dpm backups are success and most of the dpm backups consistenancy
    checks are failed.
    After investigating the log files and found two SQL server services running in the dpm 2012 r2 server those are 'sql server 2010 & sql server 2012 'service. Then i stopped sql 2010 server service and started only sql server 2012 service using (.\MICROSOFT$DPM$Acct).
    Now 'dpm console issue has gone (event id:917) but new issue ocurred 'all the scheduled job are not running' but manully i can able to run all backup without any issues. i am getting below mentioned event log errors 
    Log Name:      Application
    Source:        SQLAgent$MSDPM2012
    Date:          7/20/2014 4:00:01 AM
    Event ID:      208
    Task Category: Job Engine
    Level:         Warning
    Keywords:      Classic
    User:          N/A
    Computer:      
    Description:
    SQL Server Scheduled Job '7531f5a5-96a9-4f75-97fe-4008ad3c70a8' (0xD873C2CCAF984A4BB6C18484169007A6) - Status: Failed - Invoked on: 2014-07-20 04:00:00 - Message: The job failed.  The Job was invoked by Schedule 443 (Schedule 1).  The last step to
    run was step 1 (Default JobStep).
     Description:
    Fault bucket , type 0
    Event Name: DPMException
    Response: Not available
    Cab Id: 0
    Problem signature:
    P1: TriggerJob
    P2: 4.2.1205.0
    P3: TriggerJob.exe
    P4: 4.2.1205.0
    P5: System.UnauthorizedAccessException
    P6: System.Runtime.InteropServices.Marshal.ThrowExceptionForHRInternal
    P7: 33431035
    P8: 
    P9: 
    P10: 
    Log Name:      Application
    Source:        MSDPM
    Date:          7/20/2014 4:00:01 AM
    Event ID:      976
    Task Category: None
    Level:         Error
    Keywords:      Classic
    User:          N/A
    Computer:      
    Description:
    The description for Event ID 976 from source MSDPM cannot be found. Either the component that raises this event is not installed on your local computer or the installation is corrupted. You can install or repair the component on the local computer.
    If the event originated on another computer, the display information had to be saved with the event.
    The following information was included with the event: 
    The DPM job failed because it could not contact the DPM engine.
    Problem Details:
    <JobTriggerFailed><__System><ID>9</ID><Seq>0</Seq><TimeCreated>7/20/2014 8:00:01 AM</TimeCreated><Source>TriggerJob.cs</Source><Line>76</Line><HasError>True</HasError></__System><Tags><JobSchedule
    /></Tags></JobTriggerFailed>
    the message resource is present but the message is not found in the string/message table
    plz help me to resolve this error.
    jacob

    Hi,
    i would try to reinstall DPM
    Backup DB
    uninstall DPM
    Install DPM same Version like before
    restore DPM DB
    run dpmsync.exe -sync
    finished
    Seidl Michael | http://www.techguy.at |
    twitter.com/techguyat | facebook.com/techguyat

  • IDOC Monitoring issue - job BPM_DATA_COLLECTION* not running.

    Hi all,
    We are facing an issue with BPM "IDOC Monitoring" (under application monitoring), which we have setup to monitor Inbound and Outbound Idocs in 2 separate R/3 systems.
    In one system it works fine, and measured values are returned each time the monitor is set to run according to the specified schedule.
    However, for another R/3 system, the monitor has never run, even though the settings are identical in the monitoring setup.
    From reading the Interfaces Monitoring Setup guide, I found that this monitor depends on a job called BPM_DATA_COLLECTION* which runs in the monitored system. In the system where monitoring is functioning correctly, I can see this job is completing successfully at the time the monitor is set to run. All I find is completed jobs - no scheduled or released jobs present.
    However, for the system where the monitoring is not functioning, I found that this job is not running, but that the job is sitting in scheduled status instead.
    When I tested manually running the job in this system, it ran successfully, and in Solution Manager the monitor brought back the measured value, so the monitor only ran successfully when I manually ran the job in the monitored system.
    I read notes 1321015 & 1339657 relating to IDOC monitoring. 1321015 appears to be more relevant, yet it does not exactly describe my issue - it mentions the job BPM_DATA_COLLECTION* failing rather than just remaining in scheduled status which is what I see.
    Anyone else see this issue before?
    On a more general point - the standard BPM Setup guide doesn't really go into much detail on IDOC Monitoring, and makes no mention of what is happening in the background, i.e. the job BPM_DATA_COLLECTION* being created and run as per schedule. This info is found in a separate document "Interface Monitoring Setup Guide".
    Is there any single document which describes fully what happens both in the Solution Manager and the Monitored systems when BPM is activated? For example, to describe which monitors require jobs to be run, which monitors require additional setup in monitored system, etc? A document such as this which describes exactly the process flow for each monitor would be very useful in troubleshooting issues going forward.
    Thanks,
    John

    Hello John,
    most probably the user assigned to the corresponding RFC READ connection that connects SolMan with the backend system doesn't have proper authorization to release a job. That's why it is only created/scheduled but not released. Verify if the RFC user on the backend has the latest CSMREG profile assigned according to SAP note 455356.
    You can also check if the latest ST-PI support package is installed  on your backend system as the ST-PI usualy contain the latest definition of CSMREG.
    Best Regards
    Volker

  • Job Will Not run due to credentials

    I am attempting to set up a job to run a MSDB package stored on my server. The job is failing and giving me the error below in the history log. I have several other jobs that run MSDB packages with the same credentials and they run fine. Why would this be
    failing. I have tried using a local account, but that fails as well. I am using windows authentication and the SDC\Administrator is the main admin account on our domain.
    the job failed. It was invoked by SDC\Administrator. The last step to run was step 1.
    **Update** My job is executing with an account called Tailwind\Systems. Tailwind is the server name and consequently a user name in our sql server 2012 instance. I cannot figure out how to make it execute as another account**

    OK so start to finish.
    I have a file saved in a directory on the same server sql server 2012 is running on.
    Next I create a db and import that file in and allow the table to get created.
    Next I import again this time electing to delete existing rows and using the package do not use encryption.  I accept all defaults and save in sql server
    Now that the package has been saved I create a new job and create a step that calls this package to run it. I leave all defaults including the owner which is the Windows admin account.
    I run the job and get the error. I run the package manually and it runs. The funny thing is, this is the way I have always done it and never had this issue. I have other jobs that still successfully run packages but are in other db's. I test by creating
    a new import package and job to run but from within the already existing db and it works. So it seems to be a permission somewhere that I have yet to find.

  • Scheduled jobs do not run as expected after upgrading to 10.2.0.3 or 4

    FYI, we've had a ticket open for several days because our scheduled jobs (dbms_scheduler) would no longer run as scheduled after an upgrade to 10.2.0.4 on HPUX, couldn't find the solution by searching in metalink, nor did I find it here or using Google - obviously I wasn't searching correctly. There is a note id that references a set of steps that appears to have resolved our problem. I am putting this note out here so that if you encountered the same difficulty you may come across this note earlier in your troubleshooting efforts rather than later. The full title of the note is: 'Scheduled jobs do not run as expected after upgrading to either 10.2.0.3 or 10.2.0.4'. The Doc ID is: 731678.1.

    Thanks - our ticket should be getting closed out on it (our dba will be updating it), the scheduler has been running reliably since we took the steps in the doc mentioned.

  • Convert Spool to PDF and mail for the jobs not run in background.

    Dear All,
    I need to convert the spool job into pdf and mail as attachment for job not run in background.
    I am printing the bills  using the program  SD70AV3A to print the bills in a batch using the SUBMIT.
    But i am not using the spool options since its going to into waiting state in spool...and i need to change the printer properties to G
    to get the printer to print the reports as these jobs run in backgroud....but this solution is not acceptable solution.
    SUBMIT SD70AV3A
             WITH RG_KSCHL-LOW = 'ZRR'
             WITH RG_NACHA-LOW = '1'
             WITH PM_VERMO = '2'
             WITH PM_NSORT = '1'
             WITH RG_VBELN IN range.
    *        TO SAP-SPOOL
    *        SPOOL PARAMETERS print_parameters
    *        WITHOUT SPOOL DYNPRO
    *        VIA JOB name NUMBER number
    *        AND RETURN.
    so how do i get the spool number from the above after the job is submitted....and then convert into pdf and mail
    please suggest some solution. I have searched on net but most the solutions are say to run the job as background job...
    i have also looked into the program RSTXPDFT4 but it asks for the spool id ....how can i get spool id for the job that has been completed.
    please suggest some solution.
    Thanks
    Bhargava.

    Hi,
    Spool to PDF : FM 'CONVERT_ABAPSOOLJOB_2_PDF'
    Table TSP01 to get spool number
    You ,may need to convert OTF to PDF: FM CONVERT_OTF
    and the to binary : FM 'SMCS_XSTRING_TO_BINARY'
    To send mail use class: cl_document _bcs.
    Regards,
    Maria João Rocha

  • Cron jobs not running

    Hi
    I am using linux 5.x
    uname -r
    2.6.18-348.16.1.0.1.el5PAE
    I have 2 crontab jobs set up as following .
    cat backup_confluence.sh
    #!/bin/bash
    # bash shell script to backup confluence backups to bccpfr
    # Schedule this script after the backups are performed that day
    # Setup the environment for the bourne-again shell
    . ./.bash_profile
    # Copy the daily backups for confluence
    find /usr/share/confluence/data/backups -mtime -1 \! -type d -exec ls {} \; -exec scp {} bccpfr:/usr/backup/confluence \;
    second
    cat  backup_jira.sh
    #!/bin/bash
    # bash shell script to backup jira backups to bccpfr
    # Schedule this script after the backups are performed that day
    # Setup the environment for the bourne-again shell
    . ./.bash_profile
    # Copy the daily backups for jira
    find /usr/share/jira/backups -mtime -1 \! -type d -exec ls {} \; -exec scp {} bccpfr:/usr/backup/jira \;
    cron as following
    $ crontab -l
    13 06 * * * /home/lsg/backup_jira.sh
    18 06 * * * /home/lsg/backup_confluence.sh
    Issue
    Both the above .sh are running properly when executed  manually but they are not running automatically at the given time crontab job.
    Kindly help me on this issue.
    many thanks
    J

    Is your crontask not running at all or is the scheduled task failing?
    Hey Tommy I sent you some replies yesterday; don't know what happened to them. Yes I can see that crond runs every ten minutes. Also when I created the crontab file using the command you sent here I could see that it was running every minute by looking at root mail in /var/log.
    Do you have execute permission set on the scripts?
    Yes
    Any info in /var/log/messages or e-mail?
    Output from mail-
    From root@razorback  Tue Feb 11 08:19:01 2014
    Return-Path: <root@razorback>
    Received: from razorback (localhost [127.0.0.1])
        by razorback (8.13.8/8.13.8) with ESMTP id s1BEJ154018231
        for <root@razorback>; Tue, 11 Feb 2014 08:19:01 -0600
    Received: (from root@localhost)
        by razorback (8.13.8/8.13.8/Submit) id s1BEJ1rh018198;
        Tue, 11 Feb 2014 08:19:01 -0600
    Date: Tue, 11 Feb 2014 08:19:01 -0600
    Message-Id: <201402111419.s1BEJ1rh018198@razorback>
    From: root@razorback (Cron Daemon)
    To: root@razorback
    Subject: Cron <root@razorback> $(date) >> /tmp/cronout
    Content-Type: text/plain; charset=UTF-8
    Auto-Submitted: auto-generated
    X-Cron-Env: <SHELL=/bin/sh>
    X-Cron-Env: <HOME=/root>
    X-Cron-Env: <PATH=/usr/bin:/bin>
    X-Cron-Env: <LOGNAME=root>
    X-Cron-Env: <USER=root>
    /bin/sh: Tue: command not found
    From root@razorback  Tue Feb 11 08:20:02 2014
    Return-Path: <root@razorback>
    Received: from razorback (localhost [127.0.0.1])
        by razorback (8.13.8/8.13.8) with ESMTP id s1BEK2BD018267
        for <root@razorback>; Tue, 11 Feb 2014 08:20:02 -0600
    Received: (from root@localhost)
        by razorback (8.13.8/8.13.8/Submit) id s1BEK1mT018238;
        Tue, 11 Feb 2014 08:20:01 -0600
    Date: Tue, 11 Feb 2014 08:20:01 -0600
    Message-Id: <201402111420.s1BEK1mT018238@razorback>
    From: root@razorback (Cron Daemon)
    To: root@razorback
    Subject: Cron <root@razorback> $(date) >> /tmp/cronout
    Content-Type: text/plain; charset=UTF-8
    Auto-Submitted: auto-generated
    X-Cron-Env: <SHELL=/bin/sh>
    X-Cron-Env: <HOME=/root>
    X-Cron-Env: <PATH=/usr/bin:/bin>
    X-Cron-Env: <LOGNAME=root>
    X-Cron-Env: <USER=root>
    /bin/sh: Tue: command not found

Maybe you are looking for