How to automate full online SQL to tape/device (daily)

Dear all,
In my company, daily backups is of 2 types.
1. from Db13 SAP level backup full online to tape.
2. from SQL , full online to disk.
Db13 jobs we schedule in SAP.
But, i want to automate the SQL level full online backup to device.(External USB drive)
how to do that, if any SQL script is avaliable means, please let me know that.
thanks in advance,
by
raghav

Hello Kevin,
I want to do it from OS level (windows).
where we can create a batch file and schedule tht daily.
so, tht back up will happen to tape drive by the batch file.
i had problem in creating a batch file.
i heard tht batch file can be reated in such a way tht it will execute the same steps
that as we do it by clicking the mouse.
i.e Start>programms>ms sql > enterprise manager > and steps to navigate in enterprise manager.
if we happend to create a batch file like this.. we can schedule that from windows level.
and one more doubt:-
How to restore the BACKUP obtained from DB13 (sap level)?
how good it is when compared to SQL enterprise manager backup ?
hope you give wings to my tought.
thanking you in anticipation.
by
raghavendra rao.

Similar Messages

  • Cannot create tape device

    Hi, maybe someone can help me?
    My tape device /dev/rmt/0n was 20Gb in size even though it had no tape in it, after a failed backup.
    This was after filling up my hard-drive to 100% capacity, so I decided to delete /dev/rmt/0n and recreate using "devfsadm" or "boot -r" but neither of these commands worked!
    Now I cannot connect the tape drive at all as I cannot create /dev/rmt/0n.
    Does anybody know how this can be fixed?
    Thanks

    My tape device /dev/rmt/0n was 20Gb in size even
    though it had no tape in it, after a failed backup.Sound like you mis-specified the destination of your backup and it literally went to a file "/dev/rmt/0n" (overwriting your device file) instead of to tape.
    Please run 'drvconfig' then 'tapes'. (Or do a reconfiguration reboot.) This should recreate your device name.

  • How to automatically swap R,G,B channels on a folder full of images

    How to automate a Harris Shutter like effect to quickly apply to a few hundred images? Essentially for each finished image, I want to take the R, G, B channels from three different images. I want to do this with a folder full of images. From image #1, use the Red Channel, from image #2, use the Green Channel, from image #3, the Blue Channel. Then repeat that for all the images. Each pulling the channels from the subequent images. Any suggesions on how to automate this? Here is a tutorial I wrote up on how to do it to a single image. http://art3150.blogspot.com/2010/09/weekend-homework.html
    Another way to think of it would be like all the Red channels are original, but the Green channels would be from an image one ahead in the sequence and all the Blue channels would be from an image two ahead in the sequence. The channels would be slightly out of sync with one another.

    Something like this....
    JPGs have been assumed.
    main();
    function main(){
    var inputFolder= Folder.selectDialog ("Please select folder to process");
    if(inputFolder == null) return;
    var fileList = inputFolder.getFiles('*.jpg');
    var outPutFolder = Folder(inputFolder + "/Processed");
    if(!outPutFolder.exists) outPutFolder.create();
    while(fileList.length>2){
    var aDoc = open(fileList.shift());
    var Name = app.activeDocument.name.replace(/\.[^\.]+$/, '');
    selectChannel(0);
    dupChannel("Red");
    aDoc.close(SaveOptions.DONOTSAVECHANGES);
    aDoc = open(fileList.shift());
    selectChannel(1);
    dupChannel("Green");
    aDoc.close(SaveOptions.DONOTSAVECHANGES);
    aDoc = open(fileList.shift());
    selectChannel(2);
    dupChannel("Blue");
    aDoc.close(SaveOptions.DONOTSAVECHANGES);
    mergeToRGB();
    var saveFile = File(outPutFolder + "/" + Name + ".psd");
    SaveJPEG(saveFile,8);
    activeDocument.close(SaveOptions.DONOTSAVECHANGES);
    function SaveJPEG(saveFile, jpegQuality){
    jpgSaveOptions = new JPEGSaveOptions();
    jpgSaveOptions.embedColorProfile = true;
    jpgSaveOptions.formatOptions = FormatOptions.STANDARDBASELINE;
    jpgSaveOptions.matte = MatteType.NONE;
    jpgSaveOptions.quality = jpegQuality;
    activeDocument.saveAs(saveFile, jpgSaveOptions, true,Extension.LOWERCASE);
    function dupChannel(Name) {
    var desc4 = new ActionDescriptor();
    var desc5 = new ActionDescriptor();
    desc5.putString( charIDToTypeID('Nm  '), Name );
    desc4.putObject( charIDToTypeID('Nw  '), charIDToTypeID('Dcmn'), desc5 );
    var ref2 = new ActionReference();
    ref2.putEnumerated( charIDToTypeID('Chnl'), charIDToTypeID('Ordn'), charIDToTypeID('Trgt') );
    desc4.putReference( charIDToTypeID('Usng'), ref2 );
    executeAction( charIDToTypeID('Mk  '), desc4, DialogModes.NO );
    function selectChannel(channelNo) {
    switch(Number(channelNo)){
        case 0 : var ch = 'Rd  ';break;
        case 1 : var ch = 'Grn ';break;
        case 2 : var ch = 'Bl  ';break;
        case 3 : var ch = 'RGB ';break;
        default : break;
    var desc1 = new ActionDescriptor();
    var ref1 = new ActionReference();
    ref1.putEnumerated( charIDToTypeID('Chnl'), charIDToTypeID('Chnl'), charIDToTypeID(ch) );
    desc1.putReference( charIDToTypeID('null'), ref1 );
    executeAction( charIDToTypeID('slct'), desc1, DialogModes.NO );
    function mergeToRGB() {
    var desc3 = new ActionDescriptor();
    var list1 = new ActionList();
    var ref1 = new ActionReference();
    ref1.putName( charIDToTypeID('Dcmn'), app.documents[0].name); //Red
    list1.putReference( ref1 );
    var ref2 = new ActionReference();
    ref2.putName( charIDToTypeID('Dcmn'), app.documents[1].name ); //Green
    list1.putReference( ref2 );
    var ref3 = new ActionReference();
    ref3.putName( charIDToTypeID('Dcmn'),app.documents[2].name); //Blue
    list1.putReference( ref3 );
    desc3.putList( charIDToTypeID('null'), list1 );
    desc3.putEnumerated( charIDToTypeID('Md  '), charIDToTypeID('ClrS'), charIDToTypeID('RGBC') );
    executeAction( charIDToTypeID('MrgC'), desc3, DialogModes.NO );

  • Since the last Firefox update my computer automatically goes online at start up and I do NOT want to do this, so how can I discontinue?

    I've searched through my system preferences and Firefox preferences and both helps and cannot figure out how to prevent the computer from automatically going online (using Firefox) when I start it up. I do not want this to happen because the computer will be online when I am not using it; this poses security problems. In addition, this happened after the latest automatic update of Firefox and makes me wonder if someone is invading my privacy. I only want to go online when I open Firefox.

    Uncheck Edit > Preferences > Store > Show iTunes in the cloud purchases
    The correct way to move iTunes from one computer to another is to copy the ENTIRE iTunes folder from the old computer (or backup of the old computer) to the new computer.  This will make everything in iTunes on the new computer exactly as it was on the old computer.

  • How to automate sql server querry and output to send via email alerts

    Hi,
    I have MS SQL 2005 server. Daily i m running sql querry at 10:30 pm. i want to automate this process & output need to send via email. Please suggest what are the options availble.
    Thanks in advance
    ...Umesh

    Create a job and put the query in , then schedule it... You can email an output by using data base email feature
    EXEC msdb.dbo.sp_send_dbmail 
         @profile_name = 'Admin', 
         @recipients = '[email protected]', 
         @query = 'SELECT COUNT(*) FROM AdventureWorks.Person.Address', 
         @subject = 'Error Log Count', 
         @attach_query_result_as_file = 1 ;
    Best Regards,Uri Dimant SQL Server MVP,
    http://sqlblog.com/blogs/uri_dimant/
    MS SQL optimization: MS SQL Development and Optimization
    MS SQL Consulting:
    Large scale of database and data cleansing
    Remote DBA Services:
    Improves MS SQL Database Performance
    SQL Server Integration Services:
    Business Intelligence

  • How to automate handling of dimension members? (master data loads)

    Hi,
    We want our users to be able to upload master data (i.e. dimension member data) from a flat file. It is only possible to upload transactional data from a flat file via the Data Manager...
    Any suggestions?
    Thanks,
    Peter

    Hi Peter,
    1.  use BULK INSERT to load the flat file into SQL SERVER and automate using the SQL job or Windows schedule job. Then call the AdminTask_MakeDim package in DM. This is also be scheduled in BPC DM.
    You can also use SSIS packages to load flat file data into the SQL SERVER table.
    2. SSIS packages are not complex, you should know to use it. Basically have learnt using these SSIS packages after started working in BPC Datamanager.  You can find various online tutorials on SSIS packages.
    Flow of data should be:
    Flat file - staging table in BPC database - copy from staging tbale  to mbr<dimension> table - Admintask to process Dimension.
    The structure of staging table should be identical with mbr<dimension> table  or mbr<dimension> table should be a subset of Staging table.
    To schedule the BPC - DM packages:
    Open the eDATA menu - 'Run Package' window. You have a option of 'New automation', use that to create schedules on SSIS packages (can be used for both BPC standard packages and customized packages).
    hope it helps.
    Hi Sam,
    You can try by VBA into EXCEL, prepare the dimension member sheet and process. Basically it depends on our knowledge / ease on technology .
    I never had a chance to try on this. By visualising, this is possible but  ensure on the data type loading into the EXCEL.
    example, CompCode is 0310, while loading into EXCEL load as 310... omit the 0 in the front.
    Try and share with us how it works.
    Cheers
    Uma

  • How do i connect to sql server 2012 from cmd .

    Hi ,
    I Installed sql server 2008 R2 and 2012 Express and sql server2014 . I can conect to sql server 2008 R2 using
    SQLCMD -L command .
    How do i connect to sql server 2012 express edition from cmd .
    Any Hep appreciated
    Thanks in advance,
    Shravan

    I HAVE ANOTHER INSTANCE NAMED TEST  WHEN I USE THE COMMAND IT GIVES ME THIS ERROR 
    HOW TO RECTIFY 
    C:\Users\HP>SQLCMD -S HP-HP/TEST
    Sqlcmd: Error: Microsoft SQL Server Native Client 11.0 : Named Pipes Provider: C
    ould not open a connection to SQL Server [67]. .
    Sqlcmd: Error: Microsoft SQL Server Native Client 11.0 : Login timeout expired.
    Sqlcmd: Error: Microsoft SQL Server Native Client 11.0 : A network-related or in
    stance-specific error has occurred while establishing a connection to SQL Server
    . Server is not found or not accessible. Check if instance name is correct and i
    f SQL Server is configured to allow remote connections. For more information see
     SQL Server Books Online..
    PLZ HELP 

  • How to create a .mdf SQL Server database from a Data-Tier Application file that has data?

    This is a noob question, though I do use SQL Server databases all the time with Entity Framework when I code in C# using Visual Studio 2013.  The development environment is found below at [A].  I am trying to make a clone of a SQL Server 2008 R2
    database (.mdf)  that exists online.  I can read, connect and work with this database in Visual Studio 2013, but I wish to make a local copy of the database, as an .MDF file.  Somewhere in my notes I have a way of creating a local copy from
    an online database when using Visual Studio but I forgot how (it seems, reviewing my notes, that it deals with ADO.NET which is deprecated in Visual Studio 2013 these days, or so it seems).  So I'm looking for another way.  What I did was create
    (or export) a "Data-Tier Application File" from the online SQL Server database, with data, and it seems to have worked in that this Data-Tier Application file exists on my hard drive and seems to have data in it ("SQL Server Replication Snapshot"
    is the format it seems).  It contains skeleton code to create a database, but when I tried to execute it with SQL Server 2014 Management Studio, I got a bunch of errors.
    So my question is:
    1) Can I somehow create a .MDF SQL Server Database from an Data-Tier Application file that has data?  What tool do I use?  I saw this link, http://social.technet.microsoft.com/wiki/contents/articles/2639.how-to-use-data-tier-application-import-and-export-with-a-windows-azure-sql-database.aspx 
    and it relates to Azure, but is there a tool for C#Visual Studio 2013, standalone?
    2) If there's an easy way to create a .mdf SQL Server Database file from an online file, within SQL Server Management Studio?  I don't think so, since it would require Administrator permissions on the online server, which I don't have. I have permission
    to read, update, delete the online database file, but strangely not to download it (the service I use has a tool for backup, but not for download).
    3) same question as 2), but for Visual Studio 2013?  I don't think so, since I notice none of the templates even mentions ADO.NET anymore, but instead they go with Entity Framework.  Using EF I can of course do anything I want with the online database
    (CRUD), but it remains online.  Maybe there's a switch to make a local copy?  I guess I could write a short program to suck all the data out of the online database and put it into a new, duplicate database having the same tables, that I create on
    my localhost, but my question here is if there's an easier way than this, maybe a tool or command I can run from inside Visual Studio?
    Any advice on any of the above questions is appreciated.
    Thank you,
    Paul
    [A] Microsoft Visual Studio Professional 2013
    Version 12.0.21005.1 REL
    Microsoft .NET Framework
    Version 4.5.51641
    Microsoft Web Developer Tools 2013   2.0.40926.0
    SQL Server Data Tools   12.0.30919.1
    Microsoft SQL Server Data Tools
    Windows Azure Mobile Services Tools   1.0
    Windows Azure Mobile Services Tools

    Thanks but these links are too general to help.
    "2. what do you mean by online file?" - I mean the SQL Server database file is on a remote web server that I rent from, but I am not the administrator of.  I can access my database using SQL Server Authentication, but nothing more.
    Paul
    What do you mean by too general? It explains on how you can use data tier application to create and deploy databases
    May be this will help you to understand better
    http://www.databasejournal.com/features/mssql/article.php/3911041/Creating-Data-Tier-Applications--in-SQL-Server-2008-R2.htm
    Please Mark This As Answer if it helps to solve the issue Visakh ---------------------------- http://visakhm.blogspot.com/ https://www.facebook.com/VmBlogs

  • CPU consumption 100%: How do I see which SQL is consuming?

    Hi,
    I need help badly with this because I cannot get my head around the problem. I have a Windows server 2003 Enterprise 64-bit running Oracle 10.2.0.3 and last night it started consuming 100% CPU. The Windows performance monitor shows 100% CPU in any case, and it was difficult to even login via remote desktop.
    It is Oracle.exe that is consuming, and there is only one database on this server. It is an Opteron 875 2.2Ghz with 8GB ram.
    I want to see which SQL statement is responsible for the CPU consumption. I know exactly which user/programapplication is using it... it is useless to kick the session because the application connects immediately again from the app. server and runs the same thing. Restarting the database also has the same effect: session reconnects and CPU goes through the roof. Customer says that they are running the same stuff as always, and app. has not been updated recently. 100 % CPU to them, means something is wrong.
    Incidently, I tried to CPU patch this database a couple of days beforehand, but there was not enough space on disk (needed 1.6GB) so opatch complained and would not continue. I simply restarted the database without having run opatch. I cannot imagine that this is the cause of this problem.
    Anyhoo...
    I have a script from Kyle Haileys web site to get the CPU for each sql_id using ash tables:
    -- (c) Kyle Hailey 2007
    col type for a10
    select * from (
    select
    ash.SQL_ID , ash.SQL_PLAN_HASH_VALUE Plan_hash, aud.name type,
    sum(decode(ash.session_state,'ON CPU',1,0)) "CPU",
    sum(decode(ash.session_state,'WAITING',1,0)) -
    sum(decode(ash.session_state,'WAITING', decode(wait_class, 'User I/O',1,0),0)) "WAIT" ,
    sum(decode(ash.session_state,'WAITING', decode(wait_class, 'User I/O',1,0),0)) "IO" ,
    sum(decode(ash.session_state,'ON CPU',1,1)) "TOTAL"
    from v$active_session_history ash,
    audit_actions aud
    where SQL_ID is not NULL
    and ash.sql_opcode=aud.action
    and ash.sample_time > sysdate - &minutes /( 60*24)
    group by sql_id, SQL_PLAN_HASH_VALUE , aud.name
    order by sum(decode(session_state,'ON CPU',1,1)) desc
    ) where rownum < 10
    This give following output:
    SQL_ID PLAN_HASH TYPE CPU WAIT IO TOTAL
    cxk376gut06wn 3831248992 SELECT 12 145 5855 6012
    4ffvdt72n46yt 4094592372 SELECT 0 3 1131 1134
    dnxgm7bg8wqg1 4025412530 SELECT 3 13 337 353
    3y0dyvkc9sx69 1764840833 SELECT 18 169 56 243
    abz05715h08ct 1967470119 SELECT 0 7 175 182
    4gd6b1r53yt88 0 UNKNOWN 0 92 4 96
    9phhuzxpqwsgc 2662599208 SELECT 0 67 0 67
    275bbj5czbrwp 629013258 SELECT 0 2 64 66
    1dvrrg3aah2mu 75914095 SELECT 0 6 57 63
    My assumption is that this gives the average CPU usage over the last x minutes and the total cpu usage for this sql_id. There are no sql running that show a very high average CPU, but the total is up in 6000 for the longest running sql, so I get the statement for the top sql_id that has highest total, and it is this:
    SQL_ID cxk376gut06wn, child number 2
    SELECT NVL(MIN(ID),0) FROM TABLE_OBFUSCATED WHERE TASK = :B1 AND
    STATUS='JobQueue' AND MESSAGE_ID IS NULL
    Plan hash value: 3831248992
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | | | 3231 (100)| |
    | 1 | SORT AGGREGATE | | 1 | 42 | | |
    |* 2 | TABLE ACCESS FULL| TABLE_OBFUSCATED | 3 | 126 | 3231 (1)| 00:00:39 |
    Predicate Information (identified by operation id):
    2 - filter(("TASK"=:B1 AND "STATUS"='JobQueue' AND "MESSAGE_ID" IS NULL))
    It doesn't seem to fit! It's such a simple statement, but according to ash this is the sql using the most total cpu over the last 30 minutes?? So I create a temp index on the filter and check the plan for the statement again, and it uses the index, but I need to kill the session or bounce the database for the new plan to be used because the currently running session is still using the old plan. Customer was desperate to get it down so I even tried flushing the shared pool but new plan does not get used. I did not want to restart the database. Do I have to kill all sessions using that plan for the new one to take effect?
    Now im stuck. I don't really know how else to get an accurate picture of which sql is currently consuming CPU. I need to know this to try to fix the SQL because stopping the application is not an option, even if it is throwing too much at the database. The only other option I can think of is setting up profile limits for cpu consumption in oracle, but then the job will likely take an unacceptable amount of time for the customer.
    So my questions are:
    1. How else can I check which SQL is using CPU accurately?
    2. Does the SQL I show above accuratlely display current average CPU over x minutes with ASH? Why is it so low for the top consumer but my CPU is still at 100% on the OS?
    Thankyou for your help.

    Yeah, OK, I can see an arbitary number for CPU used since session startup (what does that refer to by the way? It can't be percent), and I can see that one user has consumed massive amounts of CPU compared to all other sessions, but he is currently not running anything, so how can I see which SQL he ran in the past day that cuased all that CPU consumption? Can I do this without AWR?
    By the way, do you need a Diagnostic Pack license for simply viewing the Performance tab of dbconsole? I Can go into dbconsole and see history for 2 days back and see the massive espike that caused CPU consumption, but when dbconsole displays the list of top SQL and top sessions under the graph, is it the top listed SQL that consumed the most cpu or do you need to correlate the colours, i.e. find the SQL listed in top SQL that has the same colour as CPU usage on the graph (green) even if it is not listed as the top "consumer" (which I assume means top AAS consumer: overall consumption of all resources)?

  • How to track current online users in OID?

    Hello,
    Is there any way to track the number of current online users in OID?(or 10Gas) therefore, I could find out who is online and how many people are online. Does OID provide the feature? Or, I have to write a Java class to record it when a user logon? what if the user doesn't logout and the seesion is timeout?
    Thanks for your help

    PL/SQL code to get number connected users since the beginning of the day :
    DECLARE
    cpt_user integer := 1;
    BEGIN
    select count(distinct(user_name)) into cpt_user from PORTAL.WWCTX_SSO_SESSION$
    WHERE (USER_NAME <> 'PUBLIC')
    AND to_char(PORTAL.WWCTX_SSO_SESSION$.LOGIN_TIME,'dd/mm/yyyy') = to_char(sysdate,'dd/mm/yyyy')
    AND IP_ADDRESS IS NOT NULL;
    htp.p(' <font color="#006699" size="1" style="font-family:Arial, Helvetica, sans-serif" > ' ||cpt_user|| ' connections</font> ') ;
    EXCEPTION
    WHEN NO_DATA_FOUND THEN
    htp.print('No connection !');
    END;

  • How to automate the queries

    hii everyone
    I want to automate queries.i.e they should be executed automatically on a particular day in the week. can we do it through oracle jobs..actually wht the case is i will be executing 5 queries evey tuesday.the output frm these queries are used to process another job tht runs every tuesday evening.rather than executing all these queries everytime i want to automate these queries.
    can anyone help me how to automate this job.
    thnakx

    hii justin
    thnkx for the reply..
    the queries wht im executing are little bit complex.one query is dependent on the other..im sending u the steps wht i will be doing when the queries are executed
    Below mentioned are the steps to be followed to check if the rewards job ran successfully or not –
    1.     Run the QUERY #1 in Rewards_Queries_Weekly_Test.sql
    2.     Note down the number of records returned in STEP 1.
    3.     Run the QUERY #2 in Rewards_Queries_Weekly_Test.sql.
    Enter current date as INPUT parameter.
    4.     Note down the number of records returned in STEP 3.
    5.     Compare the number of records returned in STEP 1 and
    number of records returned in STEP 3. BOTH THE NUMBERS
    SHOULD MATCH TO PROCEED TO NEXT STEP.
    6.     Copy the results of QUERY #1 to an excel sheet. Do not copy
    the serial number generated by PL/SQL Developer.
    Copy only the fields returned.
    7.     Copy the results of QUERY #2 to an excel sheet.
    Do not copy the serial number generated by PL/SQL Developer.
    Copy only the fields returned.
    8.     Assuming that the results of QUERY #1 were copied to the
    columns A1 & B1 and the results of QUERY #2 were copied to
    the columns D1 & E1, copy the following formula
    to F2 =IF((A2=D2)*(B2=E2),0,1)
    9.     The output of STEP 8 would be either 0 or 1 in the cell F2.
    1 indicates the records with mismatch and 0 indicates the correct
    records which can be ignored. It is possible that the records with
    mismatch have an entry in job special pay table and these records
    also can be ignored. These records can be filtered
    by following next steps.
    10.     Copy the data in the cells D1,E1 and F1 to H1, I1 and J1 respectively.
    While pasting use the option paste Special and select Values option
    in the dialog box.
    11.     Sort H1, I1 and J1 data by J1 descending and next by H1 ascending.
    This would list all the in correct records first and correct records next
    so that they can be ignored.
    12.     Now run the QUERY #3 and copy the results to K1. Do not copy the serial number generated by PL/SQL Developer. Copy only the fields returned. These records indicate the rewards special pay records given in the last one week.
    NOTE: The QUERY #3 expects two dates in the IN clause. These dates should be Current Date (Since the reward jobs run on Tuesday it will be always current Tuesday’s date) and Last Tuesday’s date.
    13.     Select all the incorrect records in the cells H1, I1 and J1 and apply
    the following conditional formatting formula with format as RED color.
    = VLOOKUP($H2,$K$2:$K$220,1,FALSE) <> " "
    NOTE: In the above formula change the value of K$220 based on the number of rows you get in the cell K from the results of QUERY #3.
    14.     The records with BLACK color are the actual records with some issue. Investigate each record manually and analyze the issue. Keep negative hours adjustment, 26 weeks rule in mind while doing so.
    the above are 14 steps i will be doing every tuesday when im executing the queries
    i want to automate the whole process wht im doing in the 14 steps..
    could u suggest me something on this
    srikanth

  • How to automate the data load process using data load file & task Scheduler

    Hi,
    I am doing Automated Process to load the data in Hyperion Planning application with the help of data_Load.bat file & Task Scheduler.
    I have created Data_Load.bat file but rest of the process i am unable complete.
    So could you help me , how to automate the data load process using Data_load.bat file & task Scheduler or what are the rest of the file is require to achieve this.
    Thanks

    To follow up on your question are you using the maxl scripts for the dataload?
    If so I have seen and issue within the batch (ex: load_data.bat) that if you do not have the full maxl script path with a batch when running it through event task scheduler the task will work but the log and/ or error file will not be created. Meaning the batch claims it ran from the task scheduler although it didn't do what you needed it to.
    If you are using maxl use this as the batch
    "essmsh C:\data\DataLoad.mxl" Or you can also use the full path for the maxl either way works. The only reason I would think that the maxl may then not work is if you do not have the batch updated to call on all the maxl PATH changes or if you need to update your environment variables to correct the essmsh command to work in a command prompt.

  • How to do profiling in sql developer?

    how to do profiling in sql developer?
    also-can somebody plz suggest a gud site for information on pl/sql code profiling.

    You might want to ask your SQL Developer-related question in the SQL Developer forum: SQL Developer
    gud site for information on pl/sql code profiling.Online Oracle Documentation:
    http://download.oracle.com/docs/cd/E11882_01/appdev.112/e10472/tuning.htm#LNPLS01214
    and/or
    http://www.oracle-base.com/articles/9i/DBMS_PROFILER.php
    Edited by: hoek on May 19, 2010 12:55 PM

  • How to publish iWeb online but not through iCloud but someone else?

    How to publish iWeb online but not through iCloud but someone else?
    Mybe Google or woodpress, 
    Or publish iWeb online with iCloud WITHOUT upgrading Lion?

    iWeb will still be around but as Jeff pointed out if you want a blog with visitors comments you should consider converting your iWeb blog to  WordPress.  Rage Software has created an applicaiton that will convert your current iWeb blog w/visitor comments to a Wordpress blog. You can get the application here: Convert Your iWeb Blog to a WordPress Blog.
    If you don't have visitor comments then the iWeb blog will continue to work for you on a 3rd party FTP server.
    If you have photo pages in your website with the accompanying slideshow you'll need to read the following:
    It's now confirmed that iWeb and iDVD have been discontinued by Apple. This is evidenced by the fact that new Macs are shipping with iLife 11 installed but without iWeb and iDVD.
    On June 30, 2012 MobileMe will be shutdown. HOWEVER, iWeb will still continue to work but without the following:
    Features No Longer Available Once MobileMe is Discontinued:
    ◼ Password protection
    ◼ Blog and photo comments
    ◼ Blog search
    ◼ Hit counter
    ◼ MobileMe Gallery
    All of these features can be replaced with 3rd party options.
    I found that if I published my site to a folder on my hard drive and then uploaded with a 3rd party FTP client subscriptions to slideshows and the RSS feed were broken.  If I published directly from iWeb to the FPT server those two features continued to work correctly.
    There's another problem and that's with iWeb's popup slideshows.  Once the MMe servers are no longer online the popup slideshow buttons will not display their images.
    Click to view full size
    However, Roddy McKay and I have figured out a way to modify existing sites with those slideshows and iWeb itself so that those images will display as expected once MobileMe servers are gone.  How to is described in this tutorial: #26 - How to Modify iWeb So Popup Slideshows Will Work After MobileMe is Discontinued.
    In addition the iLife suite of applications offered on disc is now a discontinued product and the remaining supported iApps will only be available thru the App Store from now on.
    HOWEVER, the iLife 11 boxed version that is still currently available at the online Apple Store (Store button at the top of the page) and those copies still on the shelves of retailers will include iWeb and iDVD.
    OT

  • How to automatically clear the compiler log

    Does anyone know how to automatically clear the compiler log window immediately prior to compilation? "Upon compilation" might be a better way to say it.
    I couldn't find any preference settings that would enable this functionality.
    I'm using SQL Developer version 2.1.0.63.73, which I believe is the lastest version.
    Edited by: user1983440 on Feb 11, 2010 1:21 PM

    An automatic clearing of the compiler log is not available. This would be a good enhancement request to add to the Exchange: http://htmldb.oracle.com/pls/otn/f?p=42626:16:3459578745962796
    Ashley

Maybe you are looking for