Monitoring Shell script

I want shell script for monitoring pupose (database , server ,space monitoring ) on linux , how can get the script ?
do u know the link

Ramkrishna wrote:
I want shell script for monitoring pupose (database , server ,space monitoring ) on linux , how can get the script ?
do u know the linkIf you talk about each part, there is so many monitoring events are there. if you go for space monitoring is it for trace files? archive files? FRA? ASM space? and so on..
your requirement should be clear
according to that you have to design...or you can get sample scripts from google.
Thanks

Similar Messages

  • Data guard monitoring shell script

    uname -a
    Linux DG1 2.6.18-164.el5 #1 SMP Thu Sep 3 03:28:30 EDT 2009 x86_64 x86_64 x86_64 GNU/Linux
    SQL> select * from v$version;
    BANNER
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - 64bit Production
    PL/SQL Release 11.2.0.1.0 - Production
    CORE 11.2.0.1.0 Production
    TNS for Linux: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - Production
    Hi Guys,
    I am looking for a shell script that i can cron ,which monitors dataguard env (10g and 11g )and sent email alerts if DR go out of sync say by 10 or 15 logs
    i found couple on the net but not working for some reason
    http://emrebaransel.blogspot.com/2009/07/shell-script-to-check-dataguard-status.html
    if you guys have some please share

    You are using an advanced version of Oracle and want to plug an obsolete script into it??
    Why not just monitor the Data Guard with EM or Grid Control and setup emails in there? It is 100% more reliable than anything else.

  • Shell script to monitor the data guard

    Hi,
    Can any body please provide the shell scripts to monitor the data guard in all scenarios and to get the mail when problem occurs in dataguard.
    Thanks,
    Mahipal

    Sorry Mahi. Looks like all of the scripts i've got are for logical standbys and not physical. Have a look at the link ualual posted - easy enough to knock up a script from one or more of those data dictionary views. Just had a look on metalink and there's what looks to be a good script in note 241438.1. Its a good starting point definately.
    regards,
    Mark

  • Shell script directed to one Dynamic Dashboard to monitor status of all DB

    Hi team,
    straight to scenario now..
    I have 15 databases to manage..i wrote shell scripts for monitoring each database status of ping,listener,vnc server, concurrent server, forms server, metric server, workflow, filesystem usage, alert-log etc..
    Now each and every time i will get 15 mails for one database each half n hour and it will get filled for sure if i get many alerts at single time...i thought of having a dashboard where my script output should display alerts on 3-D pie chart, bar chart etc..
    Imagine all databases statuses on one dashboard with colours displaying peaks and lows also the data gets dynamic changes every 30 mins.....
    charts will let me know to fix the issue easier..so next time i wont care how many mails reach me i will look up to dashboard and can observe what went wrong...
    please let me know any third party software available or self oracle or linux tools available.....
    hope anyone would give me suitable solution
    thanks
    dkoracle

    AFAIK Grid Control is completely free*, you just have to be careful to not go into the pages that require Management Pack licensing if you haven't purchased them for the DBs you're monitoring.
    Personally I have always used a combination of GC and shell script alerts. You don't want the GC environment to be your SPoF.
    *other than the associated server costs                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Shell script to monitor the application health deployed on weblogic

    Hi All,
    Is it possible to monitor the health of a application deployed on weblogic and send a mail if it is not running. I have JasperServer reporting application deployed on weblogic. The shell script should check if the application is running, if it down a mail should be sent.
    Thanks

    You can use WLST to get information on the state of your deployments, for example,
    the python script can be used
    class DeploymentInfo:
         def __init__(self,name,target):
              self.name = name;
              self.target = target;
         def getName(self):
              return self.name;
         def getTarget(self):
              return self.target;
    print 'CONNECT TO ADMIN SERVER';
    connect('weblogic', 'transfer11g');
    print 'OBTAINING DEPLOYMENT INFORMATION';
    deploymentsInfo = [];
    applications = cmo.getAppDeployments();
    for application in applications:
         name = application.getName();
         target = application.getTargets()[0].getName();
         deploymentsInfo.append(DeploymentInfo(name, target));
    print 'CHANGE TO DOMAIN RUNTIME ENVIRONMENT';
    domainRuntime();
    print 'APPLICATION LIFE CYCLE INFORMATION';
    applicationRuntime = cmo.getAppRuntimeStateRuntime();
    for deploymentInfo in deploymentsInfo:
         state = applicationRuntime.getCurrentState(deploymentInfo.getName(), deploymentInfo.getTarget())
         print 'Application: ' + deploymentInfo.getName() + ', State: ' + state;
         if (state != 'STATE_ACTIVE'):
              startApplication(deploymentInfo.getName());In the example above the application gets started when it is not running, you can also
    send an e-mail by using the smtplib package, for example
    import smtplib;
    server = smtplib.SMTP('email-server-host');
    server.set_debuglevel(1);
    server.sendmail(fromaddress, toaddress, message);
    server.quit();More information on the package smtplib can be found here:
    http://www.jython.org/docs/library/smtplib.html

  • Shell scripts to monitor data guard

    Hi All,
    Please help me to have the shell scripts for monitoring the data guard.
    Thanks,
    Mahi

    here is the shell script we use to monitor dataguard, it sends mail if there is a gap for more than 20 archive logs..
    #set Oracle environment for Sql*Plus
    #ORACLE_BASE=/oracle/app/oracle ; export ORACLE_BASE
    ORACLE_HOME=/oracle/app/oracle/product/10.2.0 ; export ORACLE_HOME
    ORACLE_SID=usagedb ; export ORACLE_SID
    PATH=$PATH:/oracle/app/oracle/product/10.2.0/bin
    #set working directory. script is located here..
    cd /oracle/scripts
    #Problem statemnt is constructed in message variable
    MESSAGE=""
    #hostname of the primary DB.. used in messages..
    HOST_NAME=`/usr/bin/hostname`
    #who will receive problem messages.. DBAs e-mail addresses seperated with space
    DBA_GROUP='[email protected] '
    #SQL statements to extract Data Guard info from DB
    LOCAL_ARC_SQL='select archived_seq# from V$ARCHIVE_DEST_STATUS where dest_id=1; \n exit \n'
    STBY_ARC_SQL='select archived_seq# from V$ARCHIVE_DEST_STATUS where dest_id=2; \n exit \n'
    STBY_APPLY_SQL='select applied_seq# from V$ARCHIVE_DEST_STATUS where dest_id=2; \n exit \n'
    #Get Data guard information to Unix shell variables...
    LOCAL_ARC=`echo $LOCAL_ARC_SQL | sqlplus -S / as sysdba | tail -2|head -1`
    STBY_ARC=`echo $STBY_ARC_SQL | sqlplus -S / as sysdba | tail -2|head -1`
    STBY_APPLY=`echo $STBY_APPLY_SQL | sqlplus -S / as sysdba | tail -2|head -1`
    #Allow 20 archive logs for transport and Apply latencies...
    let "STBY_ARC_MARK=${STBY_ARC}+20"
    let "STBY_APPLY_MARK= ${STBY_APPLY}+20"
    if [ $LOCAL_ARC -gt $STBY_ARC_MARK ] ; then
    MESSAGE=${MESSAGE}"$HOST_NAME Standby -log TRANSPORT- error! \n local_Arc_No=$LOCAL_ARC but stby_Arc_No=$STBY_ARC \n"
    fi
    if [ $STBY_ARC -gt $STBY_APPLY_MARK ] ; then
    MESSAGE=${MESSAGE}"$HOST_NAME Standby -log APPLY- error! \n stby_Arc_No=$STBY_ARC but stby_Apply_no=$STBY_APPLY \n"
    fi
    if [ -n "$MESSAGE" ] ; then
    MESSAGE=${MESSAGE}"\nWarning: dataguard error!!! \n .\n "
    echo $MESSAGE | mailx -s "$HOST_NAME DataGuard error" $DBA_GROUP
    fi

  • Shell script for DB backup

    HI,
    i have written some java code Database backup but there are some problems with that so now i need to write shell script for db backup.
    what i was doing in java code i was running command like that
    /usr/local/bin/tar cvzf /export/home/monitor/FILE_20091005.tar.gz FILES/*20091005.*which compress the all *20091005* files (myisam table files)
    but after compression file doesn't extract
    so i have to write shell script for that ..... can any body guide me how can i write that kind of script and put it in cron job.
    thanks

    soundar wrote:
    Hi all,
    I have migrated database from 8i to 10gr2.For Backup in 8i, we used a RMAN shell script (scheduled uding cron tab) to backup the database to Tape.(VERITAS BACKUP).
    I am new to 10G.I checked out the options to backup the database using Oracle Enterprise manager DB console.
    http://www.oracle.com/technology/obe/10gr2_db_single/ha/rman/rman_otn.htm#t1d
    I am planning to take a test backup using the steps mentioned inthe above url.Could any one suggest whcih is the best option for database backup,eiether to use Oracle Enterprise manager DB console or thru RMAN shell script for backup..?
    Edited by: soundar on Mar 9, 2010 10:53 PMDear soudar
    I woudn't suggest you to work with EM if you want to be a professional DBA. Start learning RMAN and use CLI instead of GUI
    Those who live by the GUI, die by the GUI

  • How to get ORA errors in alertlog file using shell script.

    Hi,
    Can anyone tell me how to get all ORA errors between two particular times in an alertlog file using shell script.
    Thanks

    Hi,
    You can define the alert log as an external table, and extract messages with SQL, very cool:
    http://www.dba-oracle.com/t_oracle_alert_log_sql_external_tables.htm
    If you want to write a shell script to scan the alert log, see here:
    http://www.rampant-books.com/book_2007_1_shell_scripting.htm
    #!/bin/ksh
    # log monitoring script
    # report all errors (and specific warnings) in the alert log
    # which have occurred since the date
    # and time in last_alerttime_$ORACLE_SID.txt
    # parameters:
    # 1) ORACLE_SID
    # 2) optional alert exclusion file [default = alert_logmon.excl]
    # exclude file format:
    # error_number error_number
    # error_number ...
    # i.e. a string of numbers with the ORA- and any leading zeroes that appear
    # e.g. (NB the examples are NOT normally excluded)
    # ORA-07552 ORA-08006 ORA-12819
    # ORA-01555 ORA-07553
    BASEDIR=$(dirname $0)
    if [ $# -lt 1 ]; then
    echo "usage: $(basename) ORACLE_SID [exclude file]"
    exit -1
    fi
    export ORACLE_SID=$1
    if [ ! -z "$2" ]; then
    EXCLFILE=$2
    else
    EXCLFILE=$BASEDIR/alert_logmon.excl
    fi
    LASTALERT=$BASEDIR/last_alerttime_$ORACLE_SID.txt
    if [ ! -f $EXCLFILE ]; then
    echo "alert exclusion ($EXCLFILE) file not found!"
    exit -1
    fi
    # establish alert file location
    export ORAENV_ASK=NO
    export PATH=$PATH:/usr/local/bin
    . oraenv
    DPATH=`sqlplus -s "/ as sysdba" <<!EOF
    set pages 0
    set lines 160
    set verify off
    set feedback off
    select replace(value,'?','$ORACLE_HOME')
    from v\\\$parameter
    where name = 'background_dump_dest';
    !EOF
    `
    if [ ! -d "$DPATH" ]; then
    echo "Script Error - bdump path found as $DPATH"
    exit -1
    fi
    ALOG=${DPATH}/alert_${ORACLE_SID}.log
    # now create awk file
    cat > $BASEDIR/awkfile.awk<<!EOF
    BEGIN {
    # first get excluded error list
    excldata="";
    while (getline < "$EXCLFILE" > 0)
    { excldata=excldata " " \$0; }
    print excldata
    # get time of last error
    if (getline < "$LASTALERT" < 1)
    { olddate = "00000000 00:00:00" }
    else
    { olddate=\$0; }
    errct = 0; errfound = 0;
    { if ( \$0 ~ /Sun/ || /Mon/ || /Tue/ || /Wed/ || /Thu/ || /Fri/ || /Sat/ )
    { if (dtconv(\$3, \$2, \$5, \$4) <= olddate)
    { # get next record from file
    next; # get next record from file
    # here we are now processing errors
    OLDLINE=\$0; # store date, possibly of error, or else to be discarded
    while (getline > 0)
    { if (\$0 ~ /Sun/ || /Mon/ || /Tue/ || /Wed/ || /Thu/ || /Fri/ || /Sat/ )
    { if (errfound > 0)
    { printf ("%s<BR>",OLDLINE); }
    OLDLINE = \$0; # no error, clear and start again
    errfound = 0;
    # save the date for next run
    olddate = dtconv(\$3, \$2, \$5, \$4);
    continue;
    OLDLINE = sprintf("%s<BR>%s",OLDLINE,\$0);
    if ( \$0 ~ /ORA-/ || /[Ff]uzzy/ )
    { # extract the error
    errloc=index(\$0,"ORA-")
    if (errloc > 0)
    { oraerr=substr(\$0,errloc);
    if (index(oraerr,":") < 1)
    { oraloc2=index(oraerr," ") }
    else
    { oraloc2=index(oraerr,":") }
    oraloc2=oraloc2-1;
    oraerr=substr(oraerr,1,oraloc2);
    if (index(excldata,oraerr) < 1)
    { errfound = errfound +1; }
    else # treat fuzzy as errors
    { errfound = errfound +1; }
    END {
    if (errfound > 0)
    { printf ("%s<BR>",OLDLINE); }
    print olddate > "$LASTALERT";
    function dtconv (dd, mon, yyyy, tim, sortdate) {
    mth=index("JanFebMarAprMayJunJulAugSepOctNovDec",mon);
    if (mth < 1)
    { return "00000000 00:00:00" };
    # now get month number - make to complete multiple of three and divide
    mth=(mth+2)/3;
    sortdate=sprintf("%04d%02d%02d %s",yyyy,mth,dd,tim);
    return sortdate;
    !EOF
    ERRMESS=$(nawk -f $BASEDIR/awkfile.awk $ALOG)
    ERRCT=$(echo $ERRMESS|awk 'BEGIN {RS="<BR>"} END {print NR}')
    rm $LASTALERT
    if [ $ERRCT -gt 1 ]; then
    echo "$ERRCT Errors Found \n"
    echo "$ERRMESS"|nawk 'BEGIN {FS="<BR>"}{for (i=1;NF>=i;i++) {print $i}}'
    exit 2
    fi

  • Resource utilization alerts shell scripts, Memory and CPU

    Hi All,
    I am using the Oracle Enterprise Linux verison 4.6 x64 for Oracle 10g database. I want to implement some alerts for SWAP space thresholds, CPU comsumption and memory usage thresholds. For example I want an email alert is SWAP space is used 90%. Memory usage is more than 85%. CPU is 80% busy.
    I am sure that you Gurus must have implemented this kind of alerts through some tools or shell scripts so seeking your help to implement these.
    Thanks
    Naveed

    Hi Naveed,
    Firstly, I can tell you that Oracle has its own SNMP MIBs which can be used by SNMP to send trap (alerts) to various SNMP monitoring systems. If however, you are a bit like myself and perhaps don't have the cash or the need to use something so fancy, you can use Shell Scripting or in my case something like PERL.
    By using commands such as 'grep', 'awk' and 'less' (to name but a few) on can monitor simple things such as diskspace or memory.
    For example, if I were to look for diskspace usage on a partition known as u01 I could pipe in the output of 'df -h' to PERL, then run a regular expression on it looking for anything greater than 80%.
    Example of PERL code;
    # read in df -h command, looking for the partition /u01
    open(DISKUSAGE, "df -h | grep u01|");
    $diskUse = <DISKUSAGE>; #read its input
    chomp $diskUse; #remove last characterThis reads in the command into a 'buffer' in which you can search for things, i.e. to see if the disk usage is equal to, or greater than 80% from the variable $diskUse. As a starter I would look at splitting the string on empty spaces into an array for example, and reading the array index containing your disk usage. This can also be done with a Shell Script.
    Perl also has a very easy to use SMTP module which can be used for sending E-Mails. Do a search for Net::SMTP on google on how to use it. It really is very simple.
    I'm not expert on this stuff, and I'm sure there are many other and better ways of doing such things, but I hope this at least points you in the right direction.
    Best of luck,
    Conor

  • Which is better to use Shell Script or Alerts in OMS

    HI
    This is a point of confusion for me to decide that for monitoring the databases what should be used the Alerts Set in the Management Server or Shell scripts .
    May I have your valuable suggestions with reason for supporting the suggestion

    On the other hand, if you want is the basic/simple stuff (which should be true for most cases), I woule recommend using the standard alerts coming with OEM. One less thing to maintain. Though not pretty, but usable.

  • BPEL processes and shell scripts

    Hi,
    I wish to invoke BPEL processes from a shell script so that it could be regulated from a typical Control-M environment. Can anybody let me know if its possible...?
    Regards,
    Roy

    Some kind-of Java interface would just be fantastic. Is there anything like that...?
    Well, I intend to execute some batch processes which would be scheduled and administered by Control-M; If i opt to use ws-clients within the same then auditing and monitering the status of my executing process would be an obvious overhead that i might need to hard wire in my implementation. Therefore, should it be possible to execute a BPEL process from a Java like interface, which i'd then wrap in a shell script; then Control-M on itself would be capable of monitoring the process status and all other performance parameters around the same.
    Suresh - Thanks for your suggestion, will definately post the message on those groups as well.

  • Screencapture shell script

    I am trying to capture the front most window and running this simple screencapture shell script works fine except that it requires user interaction.  The user must still mouse click on the window to complete the capture.  I know there must be a way to tell the shell script to perform without requiring user action, but I have not been able to discover how to make the script run independent of user action.  Any suggestions are appreciated.  Thanks.
    do shell script "screencapture  -w /Users/admin/Desktop/`date +%b%d%H:%M:%S`.png"

    Maybe this will help. If you monitor the "More Like This" box (top right), other threads appear. Opening them usually displays other threads.
    https://discussions.apple.com/message/6731129#6731129

  • Shell script for mail conf

    Hi all,
    DB oracle 10g
    solaris,
    It is new env ....we are not having grid configuration ....
    Please any one provide to me a script for monitoring threshold limits like >90% critcal alert ....I have written a sql script but I am unable write shell script .......my requirement is I want to call sql script (tbsthershold.sql) these alerts are configure to mail to dba group through shell script when ever tablespace reaches thershold....
    please help me out if any one having any shell scripts.... for that
    thanks,
    kk

    Hi all,
    DB oracle 10g
    solaris,
    It is new env ....we are not having grid configuration ....
    Please any one provide to me a script for monitoring threshold limits like >90% critcal alert ....I have written a sql script but I am unable write shell script .......my requirement is I want to call sql script (tbsthershold.sql) these alerts are configure to mail to dba group through shell script when ever tablespace reaches thershold....
    please help me out if any one having any shell scripts.... for that
    thanks,
    kk

  • Shell script to logout?

    Hello, I am working on an automator app that will quit all apps (which works very well), pause for several seconds and then logout a user. Right now, I am having to use the "watch me" to log out, but, it is not totally consistent in the action. Is there a shell script command?
    Thanks

    So, call me a picky, 'my system has to be just so' kinda guy... But I've got 5 users on my iMac. The wife and three kids. The kids have time limits on their accounts....
    Here's the thing with the time limits parental controls. It doesn't actually log out the user. (I guess that's to prevent lost documents etc.)
    Sounds nice, right? Well... check this out. My kids have parental controls on cause they REALLY like to play world of warcraft. For HOURS. So... WOW has this nice feature that doesn't let you exit the program nicely when you're in 'combat'. So, if you're in combat, and you run out of time with parental controls, the computer doesn't put the user back to the login screen. My kids have figured this out and will play for 15 hours at a stretch if we let them. (I think they should be allowed to play for a little extra time for figuring out how to beat the parental controls, but the wife disagrees)
    I remember way back when with 10.5.0 WOW and Parental controls didn't get along. -There's gotta be a relationship to this behavior.
    Anyways... Back to the problem. I'm at the login screen and no password for the kids account. Assume I can't log them in to log them out. (there's 4 users anyway - it would take too long)
    Okay then... Assume I'm logged into my user account (an admin account on the computer) and I see that there's 5 users logged in. Myself, my wife, and my 3 kids by typing users or w in the terminal. The parental controls have left my users logged in. So I pull off a killall -u username command in the terminal. I've found that this doesn't actually log out users, it just kills all their processes.
    When I type users or w in the terminal it shows me that there's indeed still 5 users logged in. The activity monitor as well tells me that there's no processes active that belong to anyone but me, and root. Yet the shell tells me that there are still users active.
    I KNOW there's a unix command that will log out a user, but no one will share this secret! It's not anywhere on the web. I've been looking for weeks!
    The reason I know, is that there's a command in Apple remote desktop to force users to log out.... Not only that, but it does it in a way that doesn't kill the loginwindow or the WindowServer processes.
    Someone PLEASE type the wee command that it takes to do this seemingly 'easy as two clicks' task.
    -Graham

  • Executing shell script using OSLinetoken fetchlet

    Hii,
    I do have a requirement. I need to use a shell script in the OSLineToken fetchlet. In response metric i will be checking whether the directory exists or not on the server. In order to check the existence of the directory, i have created a shell script. But how can i relate its result with the Response metric? The shell script is as follows:
    Shell Script:
    if test -d $1 ; then
    echo "DIR exist"
    else
    echo "false"
    fi
    The Response metric for the same will be:
    <QueryDescriptor FETCHLET_ID="OSLineToken">
         <Property NAME="command" SCOPE="GLOBAL">
              sh {dir_name where the shell script is uploaded}/{shell script file name} {dir_name_parameter} </Property>
         <Property NAME="startsWith" SCOPE="GLOBAL">em_result=</Property>
         <Property NAME="delimiter" SCOPE="GLOBAL">|</Property>
    </QueryDescriptor>
    Please suggest what is the use of em_result here?
    once the existence of the directory is checked, if it up then i need to call another shell script in order concatenate the contents of all the files with extension .log(this will be the parameter of shell script). Get the output from shell script and display it into custom management plug-in. As i am using cat *.log>>consolidatefile command to concatenate the data, i need to read consolidatefile file from the server and return this concatenated file data into plug-in. Again, how can i read the content of consolidatefile file in EMF? I will be creating another matric for this purpose say "read_content". the querydesciptor of the same will be as follows:
    <QueryDescriptor FETCHLET_ID="OSLineToken">
         <Property NAME="command" SCOPE="GLOBAL">
              sh {dir_name where the shell script is uploaded}/{shell script file name} {dir_name_parameter} {extension of the files to concatenated} </Property>
         <Property NAME="startsWith" SCOPE="GLOBAL">em_result=</Property>
         <Property NAME="delimiter" SCOPE="GLOBAL">|</Property>
    </QueryDescriptor>
    I am not sure which all properties to be used in this case..I have seen multiple sample files some of them uses perbin, scriptsdir but some of them does not..The related pdf also does not say anyhting about such kind of properties. Please suggest.
    I hope the explaination of the problem is not so cumbersome. Please let me know if you have any query to understand.
    Thanks,
    AS

    If you notice, localScriptsDir is a directory within scriptsDir. If you package your plug-in up and deploy it through the UI, any scripts you create will go into %scriptsDir%/emx/<target_type>. So localScriptsDir just specifies that directory for you. You don't need it but then in the command paramater you'll have something like:
    sh %scriptsDir%/emx/yourtargettype/yourscript...
    So whether you specify it in the command or another property (localScriptsDir) doesn't really matter.
    You can create your own properties in the QueryDescriptor. Just make sure you have the correct scope specified and it should be fine (options for scope are described in the Enterprise Manager DTD section of the Extensibility Guide).
    Metric collection isn't really meant for dynamic specification of input parameters. I can think of a few solutions:
    1) Create a target instance for each log directory. When you create the instance, the directory is specified. If you need to monitor a different directory, you can just create another instance. Upside is that it's flexible and scalable, and also, when you get an error you'll know exactly which directory it is based on which instance throws the error. Downside is that you have to have a separate instance for each directory.
    2) If the log directories are well known and finite (and won't change names), hardcode them into the target metadata. Have a different metric collect for each log directory, so you'll have as many metrics as log directories you want to monitor. Even if the names of the directories are different, you can use instance properties to map them, so if you know there will always be 5 log directories you want to monitor, you can have 5 instance properties to map the names into the metrics, although this won't work if you don't have the same number each time. Upside is that there is only a single target instance. Downside is that it's not as flexible.
    3) Use a job rather than a target type to find out this information. You could create a new job type which scans the logs for information and have the directory as an input parameter to the job. You could have this job on a repeating schedule to duplicate the effect you are trying to get out of creating a target type. The upside is that you can start the job whenever you want from the UI and specify exactly which directory whenever you run it. The downside is that the job system is centered on the OMS rather than the agent, so every time it runs it will have to contact the agent to do the work. In the case of the target type, the agent acts autonomously without contact from the OMS.
    There are probably other options, but these are the quick ones off the top of my head.
    Chris

Maybe you are looking for

  • Problem converting XML back to ABAP Structure using XSLT program

    Hi Experts, I am currently trying to convert this xml file back to my ABAP structure but i have been unsuccessful so far. In XML file contain more than 1 reocrd. But my XSLT program is retriving only only first record. Try to help to get all the reco

  • Composite foreign key issue

    Hi, I am trying to create a composite foriegn key. I have created my primary key as a composite key, made up of three fields. Table 1 PatientID - Made from 3 fields(a,b,c)(primry key) Table B PatiendID- (should be the foreign key which references tab

  • How to add payment advice for XML file filed in vendor account group

    Hi All, I have a requirment to add Payment advice for XML file field in vendor account group under payment transcation tap, kindly advice where i can add above mention field in vendor account group. thanks khaja

  • Links to screenshots in a pdf

    We publish our documents in pdf, but we use FrameMaker to create our documents, and we create cross-references for all of our screenshot captions.  This is done in case the screenshot is moved to another page by an edit and to automatically update th

  • ITunes Used to Work, After Recent Update, itl file error

    I have 2 huge need questions to be answered. 1. First off, my iTunes worked great. Never was anything wrong, went online at my work and downloaded the lastest update to a cd. Brought the cd home and installed it onto my computer. Then I click on the