Program executing faster than OS level calls... wait()/sleep()/etc...?

I have a class that uses makes an OS level call to move a file from location A to location B, and then I have further code that will open this file and then do some reading and writing to it.
The problem I am having is that the OS command is issued and then it tries to open the file for processing... however, my code is trying to open the file and process it before the OS has moved it... i.e. I am trying to open a file in location B before the move command has fully finished. Is there a way to deal with this?
Problem 2... is similar...
This time I am writing a file and then I close it and then I issue an OS level call that will move the file to some location... The problem is that the OS is trying to move the file before it has closed... Once again how can I deal with this?
I was trying to use the wait() method of the Object class however, this does not appear applicable to my situation as it is not able to fully determine how far along my OS level calls are.
I know that the situation is true because if I put in a sleep function for some time then everything works fine, however sleeping is not the most desired way for one of 2 reasons... one the sleep time may not be long enough and thus cause the same problem as above... or if the sleep time is to long then I am wasting uneccessary time sleeping.. and in effect doing nothing.
Any thoughts?
Problem 1 pseudocode...
Process p = Runtime.getRuntime().exec("mv file1 file2");
//open file2
//cant because the mv is still taking place.
// if i use a Thread.sleep(1000) then all is fine.. however, this is not desired.
fw.write("some text");
fw.close();
Process p = Runtime.getRuntime().exec("mv file1 file2");
//cant because either the file is still being written tooo or file is not closed yet.
// if i use a Thread.sleep(1000) then all is fine.. however, this is not desired.
Any thoughts or help/pseudocode would be greatly appreciated.
Thanks in advance.
echardrd

Thanks that will make sense for where I have an OS level call and then I need it to wait before I can read the file...
that is
Process p = Runtime.getRuntime().exec("mv file1 file2");
//new code
p.waitFor();
// code to open and read file...
however how about the inverse?
That is my other problem... which is where i have java code that writes a file and after the file has been written... then have an OS level call to move the file...
i.e.
fw.write("some text");
fw.close();
//OS level call.... for example
Process p = Runtime.getRuntime().exec("mv file1 file2");
what I am getting at is the file is still open before my OS level call is issued... the waitFor() will work for the previous problem.. but is there a way to do a wait so that my file has time to close before I issue my OS level call...? without using a sleep.?
Once again thanks in advance...
echardrd

Similar Messages

  • Performance... Why a function column in a view is executed more than once...?

    Why a function column created inside a view is executed more than once when called more than once?
    EXAMPLE:
    create or replace view aux1 as
    date_column,
    any_function(date_column) column1
    from any_table
    create or replace view aux2 as
    column1 c1,
    column1 c2,
    column1 c3
    from aux1
    select * from aux2
    It will execute 3 times the function any_function... logically the value will be the same for all columns...
    I understand why!... are 3 calls... but...
    Why not to create a "small" verification and if the function column was execute replace the second, the third... value? ... instead of execute 3, 4... times...
    tks
    Braga

    Actually, this is more than a performance issue. This is a consistency problem. If the function is NOT deterministic then you may get different values for each call which is clearly not consistent with selecting 3 copies of the same column from a row. Oracle appears to have fixed this in 9i...
    Connected to:
    Oracle8i Enterprise Edition Release 8.1.7.2.0 - Production
    With the Partitioning option
    JServer Release 8.1.7.2.0 - Production
    create view v1 as select dbms_random.value(1,100) r from dual;
    create view v2 as select r r1, r r2 from v1;
    select * from v2;
              R1           R2
              93           74
    Connected to:
    Oracle9i Enterprise Edition Release 9.0.1.3.0 - Production
    With the Partitioning option
    JServer Release 9.0.1.3.0 - Production
    create view v1 as select dbms_random.value(1,100) r from dual;
    create view v2 as select r r1, r r2 from v1;
    select * from v2;
              R1           R2
              78           78Richard

  • Any Program to sort by keywords faster than CS-PS6-Bridge For Keyword Search?

    To sort by keywords in (CS-PS6)-Bridge  one has to group all your folders under one overall folder. Then when you search by keywords (CS-PS6)-Bridge opens up all the photos in Bridge which take  avery long time. Since Lightroom Smart Previews open up smaller files will it open up all ones pictures faster? and then can it sort by keywords.
    I am assuming Lightroom  smart preview opens up a larger size file than CS6-PS-6-Bridge and thus is even slower.
    So the question becomes are there Any Program to sort by keywords faster than CS-6-Bridge For Keyword Search? i.e A program that doesn't have to open up a file to look at keywords or a way to open smaller thumbnails in Bridge?

    Really?
    You mean you can't just click on the Magnifying Glass in the Search Box like I can?
    Uploaded with plasq's Skitch!
    And then Select Rating?
    Uploaded with plasq's Skitch!
    And then click on the Third Dot and it will become a Star?
    Uploaded with plasq's Skitch!
    Or even use a Smart Album? File -> new Smart Album: My Rating -> is -> And there's click on the third dot to make it a star trick again?
    Regards
    TD

  • Call waiting says "multiple calls" rather than listing the number or the contact.  Does anybody know how to fix this?

    Call waiting says "multiple calls" rather than listing the number or the contact.  This error is not consistent, sometimes it says "multiple calls" and other times it lists the contact.  This is happening on a coworker's iPhone 4s.  Does anyone know how to get it to list the contact every time?
    Thank you,
    Dan

    It sounds as though it may not be ending calls correctly, I would try restarting the phone, and then resetting all settings.

  • PowerMac faster than intel Mac when running video programs?

    Hello everyone! Having a question now so wonder if any of you have the same experience?
    Recently i have many video works to do. I have a powerbook, which is the lasted 17inch powerbook and an intel imac (early 2008, 2.66GHz). I run the same version of finalcut pro and compressor. But thing become very strange, which is that when i run compressor to make final video files, my powerbook is much faster (about 2 times faster) than intel iMac. When i run XDCAM program to import video clips to finalcut from Sony XDCAM blueray reader, powerbook is also much more faster than intel iMac, and which is about 10time more faster than intel iMac. So it become very strange. I list my configurations below, and any of you can help me to solve the 'problem' or explain the strange fenominal.
    Maybe theres anything wrong with software configurations, any one can also help me with that?
    Thank you very much!
    Powerbook G4 17inch:
    CPU: powerpc G4 1,67GHz
    RAM: 2GB (two 1GB module)
    HDD: 120GB IDE
    System: Mac OS X Leopard 10.5.8
    Program: Final Cut Studio 2
    iMac 20inch Early 2008
    CPU: Intel Core 2 Duo 2,66GHz
    RAM: 4GB (two 2GB module)
    HDD: Cosair 256GB SSD
    System: Mac OS X Lion 10.7 (tried also on Snow Leopard)
    Program: Final Cut Studio 2 (tried also Final Cut Studio 3 on friends imac with the same configuration)
    Camera:
    Sony XDCAM PDW530 Broadcast Camcorder.
    Media:
    Sony XDCAM blueray disk
    Reader:
    Sony XDCAM blueray disc drive
    Program:
    XDCAM transfer for Final Cut Pro 2.12
    XDCAM browser software 1.20

    Of course it depends on what you're doing and how well you know how to use available resources... the general rule for Logic is:
    >>>Fastest machine with the most RAM you can afford.<<<
    That said, someone who knows what they're doing can get an amazing amount of effects and virtual instruments on an old PPC G5 computer running Logic 8.
    So, while fast machines will allow more plugins/VI with less overloads... knowing what you're doing will take you further so that the difference between machines would not make a difference in the music.
    pancenter-

  • Threads to keep the CPU faster than the disk?

    Greetings,
    I hope this is the correct place to post a question like this - I did not see a forum specific to C/C++ programming on Solaris.
    I'm currently writing a data conversion program and would like to get better performance. The programming language is straight C, Solaris 10, E25K with 8 CPUs and 16G RAM allotted to my zone. I do not have admin on the box. I am compiling 64-bit with lots of compiler options for performance.
    The process is very linear and most of the optimization examples I find are for making loops run in parallel and such. Well, I don't have any loops. I'm moving a lot of data from a set of source files, doing some transformation and validation, then writing to the appropriate target file. No recursion or matrix math here...
    I wrote my initial test program which would basically spin through the source files and write empty target files. Doing this I was able to process about 70,000 source records per second - which was acceptable and on par with the speed of simply copying the disk files from one place to another.
    Once I started adding logic, the records per second started to drop drastically. I expected this to some degree, but adding just the basic initial logic cut the records per second in half, and after that the performance dropped in a pretty linear fashion as I added transformation logic. Mind you, most of the logic is moving source to target and space padding the target, validating a date range, etc.. Nothing complex by any stretch of the imagination, there are just a lot of fields.
    Before I spend a lot of time trying to multi-thread the application, I wanted to see if my expectations are realistic. My thinking is that 8 CPUs should be able to keep up with the disk subsystem and that my conversion should not take any longer than the amount of time it takes to simply copy the data from one point to another. Possible?
    Currently I'm processing like this:
    1. mmap open all sources (there are about 10 to 15 depending)
    2. collect counts of all source records in a given "set"
    3. wait for any previous targets to finish writing to disk
    4. process the current set of source records and write target records to memory buffers for each target
    5. when a given target buffer is full, aiowrite to the target file
    6. while there are source records, goto step 2
    Basically I used aiowrite to get a little free async operation in that any target buffers that are ready to be written could do so while the next set of source records is being grouped (being read from the mmap'd source files). I also try to keep things as fast as possible by not moving the data more than necessary. Usually my transformation logic can move the data directly from the mmap'd file to the target buffer, and in other cases only a single move of the data needs to be done.
    What I think I would like to do is create a thread that groups the source record sets into 8 independent memory locations. This thread's job is to simply keep those group locations full. Then 8 worker threads would pick the next source "set" from the pool and process it, and only have to sync on a mutex when writing to the target file.
    Any insight or feedback would be greatly appreciated.
    Thanks,
    Matthew

    If your application is not using threads, then the entire program is running in a serial state, waiting for i/o etc. To check how your little program works, use:
    truss -a -d -D -f -l -o your_truss_output_file.log your_application

  • Web Report - ABAP Vs JAVA engine - ABAP 10 times faster than JAVA

    Guys,
    I want to share what we found in our project and see if any of you have insights
    into our findings.We are on NW2004S SP14 and we are moving to SP15 in a couple of weeks.We created query, developed WAD for it and executing the WAD takes for this query takes 22 secs (Vs 2 secs using ABAP) the query output has 1 million records and most of the actions we take from that point on like right click on account takes 20 secs (Vs 0 secs/instant using ABAP) , drilldown to level 4 of account hierarchy takes 58 secs (Vs 5 secs using ABAP), drilldown on cost center level 6 takes 42 secs (Vs 4 secs using ABAP), , right click on cost center takes 32 secs (Vs 3 secs using ABAP), ..etc.
    Basically every action we take in the JAVA report takes an average of  28 secs.There are 9 aggregates built on the cube that are barely hit by this query but the same query performing same actions with same selections hit the aggregates many many times.The questions I have is why is ABAP so fast compared to JAVA ? What is true explanation behind this behavior ? What are the dis-advantages by using ABAP engine ? Users are loving the performance and features of ABAP while they weren't really on board with the original JAVA report (as it was slow). ABAP is sure enough 10 times faster than JAVA. Query/Query Properties are exactly the same in ABAP and JAVA.Please explain.
    Cheers
    RT

    Hi All,
    Thanks to all you for your responses. I appreciate your time for going through my questions and coming forward to express your views.
    However, I was looking for more specific "factual" answers. My question is "What does a client miss if they opt to install only ABAP based BI 7.0, as against JAVA Based BI 7.0"
    thanks again.
    Naga

  • (  name-service call wait   ) event   is amoung the top 5 wait events

    10.2.0.3 2node- RAC
    When I check the wait event in the active session wait event the sessions can be see that they are blocked by the LMON process. The wait event appears only on one node.
    It is very hard to find information on the net or on the metalink.
    How could I overcome this wait event ?
    select program, type from v$session
         where sid in
         (select blocking_session from v$active_session_history
                   where event like 'name-service%'
         and rownum < 1000 )
    oracle@dbokyanus1 (LMON) , BACKGROUND
    Top 5 Timed Events Avg %Total
    ~~~~~~~~~~~~~~~~~~ wait Call
    Event Waits Time (s) (ms) Time
    db file sequential read 4,331,507 15,218 *4* *31.2*
    CPU time *8* *17.5*
    log file sync 1,462,968 5,513 *4* *11.3*
    name-service call wait 72,058 4,545 *63* *9.3*
    SQL*Net message from dblink 4,197 4,047 *964* *8.3*

    oceanic-815 wrote:
    10.2.0.3 2node- RAC
    When I check the wait event in the active session wait event the sessions can be see that they are blocked by the LMON process. The wait event appears only on one node.
    It is very hard to find information on the net or on the metalink.
    If I use Google to find information about this
    I ran into this page.
    There is indeed not much to find about this wait-event, other than there is probably some sort of network communication problem between the two nodes.
    Maybe the best solution is to raise a S/R at Oracle support

  • Is this logging code faster than using a standard logging API like log4J

    is this logging code faster than using a standard logging API like log4J or the logging API in java 1.4
    As you can see my needs are extremely simple. write some stuff to text file and write some stuff to dos window.
    I am thinking about using this with a multi threaded app. So all the threads ~ 200 will be using this simultaneously.
    * Tracer.class logs items according to the following criteria:
    * 2 = goes to text file Crawler_log.txt
    * 1 = goes to console window because it is higher priority.
    * @author Stephen
    * @version 1.0
    * @since June 2002
    import java.io.*;
    import java.net.*;
    import java.util.*;
    import java.text.*;
    class Tracer{
    public static void log(int traceLevel, String message, Object value)
    if(traceLevel == 1){
    System.out.println(getLogFileDate(new Date()) +" >" + message+ " value = " + value.toString()););
    }else{
    pout.write(getLogFileDate(new Date()) +" >" + message + " value = " + value.toString());
    pout.flush();
    public static void log(int traceLevel, String message )
    if(traceLevel == 1){System.out.println(message);
    }else{
    pout.write(message ) ;
    pout.flush();
    //public static accessor method
    public static Tracer getTracerInstance()
    return tracerInstance;
    private static String getLogFileDate(Date d )
    String s = df.format(d);
    String s1= s.replace(',','-');
    String s2= s1.replace(' ','-');
    String s3= s2.replace(':','.');
    System.out.println("getLogFileDate() = " + s3 ) ;
    return s3;
    //private instance
    private Tracer(){
    System.out.println("Tracer constructor works");
    df = DateFormat.getDateTimeInstance(DateFormat.MEDIUM, DateFormat.MEDIUM);
    date = new java.util.Date();
    try{
    pout = new PrintWriter(new BufferedWriter(new FileWriter("Crawler_log"+getLogFileDate(new Date())+".txt", true)));
    pout.write("**************** New Log File Created "+ getLogFileDate(new Date()) +"****************");
    pout.flush();
    }catch (IOException e){
    System.out.println("**********THERE WAS A CRITICAL ERROR GETTING TRACER SINGLETON INITIALIZED. APPLICATION WILL STOP EXECUTION. ******* ");
    public static void main(String[] argz){
    System.out.println("main method starts ");
    Tracer tt = Tracer.getTracerInstance();
    System.out.println("main method successfully gets Tracer instance tt. "+ tt.toString());
    //the next method is where it fails - on pout.write() of log method. Why ?
    tt.log(1, "HIGH PRIORITY");
    System.out.println("main method ends ");
    //private static reference
    private static Tracer tracerInstance = new Tracer();
    private static Date date = null;
    private static PrintWriter pout = null;
    public static DateFormat df = null;
    }

    In general I'd guess that a small, custom thing will be faster than a large, generic thing with a lot of options. That is, unless the writer of the small program have done something stupid, og the writer of the large program have done something very smart.
    One problem with java in this respect is that it is next to impossible to judge exactly how much machine-level processing a single java statement takes. Things like JIT compilers makes it even harder.
    In the end, there is really only one way to find out: Test it.

  • Why is kernel-2.6.9 (OEL-4) faster than kernel-2.6.18 (OEL-5) ?

    Hi,
    as long as RHEL-5 and then OEL-5 have been released, I have been wondering why my own programs, compiled and run on RHEL-5/OEL-5, are slower than the same programs compiled and run on RHEL-4/OEL-4 on the same machine. This is really barmy since gcc-4.1, shipped with RHEL-5/OEL-5, is very aggressive compiler and produces faster binary code than gcc-3.4.6, shipped with RHEL-4/OEL-4. I verified this hundred times testing both compilers on RHEL-4/OEL-4 and RHEL-5/OEL-5. The 4.1 compiler always produces faster executable on the same OS.
    The problem is obviously in kernel-2.6.18. There is something in the kernel (maybe scheduler?) that slows down the execution of programs. But what? I experimented with changing various kernel boot parameters (eg "acpi=off" etc), even tried to recompile the kernel many times with various combinations of config parameters, and nothing helps. Thus, I'm still wondering whether the problem is solvable by disabling one or more config parameters and recompiling the kernel, or is deeply embedded in the main kernel code.
    Is there anybody in this forum who experienced the same, say running OEL-4 before migrating to OEL-5?
    Here are two examples showing different execution times on OEL-4.5 (kernel-2.6.9-55.0.5.0.1.EL.i686, gcc-3.4.6-8.0.1) and OEL-5 (kernel-2.6.18-8.1.10.0.1.el5, gcc-4.1.1-52.el5.2). The first example is trivial but very sensitive to overal system load and kernel version. The second example is "Sieve of Eratosthenes" - the program for finding prime numbers (CPU bound).
    EXAMPLE 1.
    /*  Simle program for text screen console  */
    /*  very sensitive to overall system load  */
    /*  and kernel version                     */
    #include <stdio.h>
    int main(void)
        register int i;
        for(i = 0; i < 1000000; i++)
         printf(" %d ", i);
        return 0;
    /* end of program */
    $ gcc -O2 -o example1 -s example1.c
    $ time ./example1The average execution times on OEL-4.5 and OEL-5 are as follow:
    Mode      OEL-4.5         OEL-5
    real      0m3.141s        0m4.931s
    user      0m0.394s        0m0.366s
    sys       0m2.747s        0m4.563s
    ----------------------------------As we can see, the program on the same machine, compiled and run on OEL-4.5 (gcc-3.4.6 and kernel-2.6.9) is 57% faster than the same program compiled and run on OEL-5 (gcc-4.1.1 and kernel-2.6.18), although gcc-4.1.1 produces much faster binary code. Since the times the process spent in user mode are almost equal on both OS, the whole difference is due to the time the process spent in kernel mode. Note that kernel mode (sys) is taking 66% more time on OEL-5. It tells me that "something" in the kernel-2.6.18 slows down the execution of the program.
    In the second example OEL-4.5 is also faster than OEL-5, but the differences in execution times are not so drastic as in the first example.
    EXAMPLE 2.
    /*           Sieve of Eratosthenes           */
    #define GNUSOURCE
    #include <stdio.h>
    #include <stdlib.h>
    #define MAX_PRIME_AREA 100000
    #define REPEAT_LOOP 10000
    int main(void)
        int prime, composite, count;
        char *sieve_array;
        if ((sieve_array = (char *) malloc( (size_t) (MAX_PRIME_AREA + 1))) == NULL)
         fprintf(stderr,"Memory block too big!\nMemory allocation failed!\a\n");
         exit(EXIT_FAILURE);
        for(count = 0; count < REPEAT_LOOP; count++)
         for(prime = 0; prime < (MAX_PRIME_AREA + 1); prime++)
                 *(sieve_array + prime) = (char) '\0';
         for(prime = 3; prime < (MAX_PRIME_AREA + 1); prime += 2)
             if (! *(sieve_array + prime) )
              *(sieve_array + prime) = (char) 'P';  /* offset prime is a prime */
                 for(composite = (2 * prime); composite < (MAX_PRIME_AREA + 1); composite += prime)
                  *(sieve_array + composite) = (char) 'X';  /* offset composite is a composite */
            /* DO NOT COMPILE FOR TEST !!!
            fprintf(stdout, "\n%d\n", 2);
            for(prime = 3; prime < (MAX_PRIME_AREA + 1); prime += 2)
                if ( *(sieve_array + prime) == 'P' )
                    fprintf(stdout, "%d\n", prime);
        free(sieve_array);     
        return 0;
    /* End of Sieve of Eratosthenes */The average execution times on the same machine on OEL-4.5 and OEL-5 are:
    MAX_PRIME_AREA     Mode         OEL-4.5         OEL-5     
                       real         0m9.196s        0m10.531s
       100000          user         0m9.189s        0m10.478s
                       sys          0m0.002s        0m0.010s
                       real         0m20.264s       0m21.532s
       200000          user         0m20.233s       0m21.490s
                       sys          0m0.020s        0m0.025s
                       real         0m30.722s       0m33.502s
       300000          user         0m30.684s       0m33.456s 
                       sys          0m0.024s        0m0.032s
                       real         1m10.163s       1m15.215s
       400000          user         1m10.087s       1m14.704s
                       sys          0m0.075s        0m0.079s
    ---------------------------------------------------------Does this ring a bell with anyone? Any clue why?
    N.J.

    An hour? Hard to believe or is your hardware that
    old?An hour? That's a super good time for 3 kernel
    packages (i686, xen and PAE) with all modules, plus 3
    kernel-devel packages, plus debuginfo package of
    150-580 MB where smart people at Red Hat decided to
    put uncompressed vmlinux image which is necessary for
    kernel profiling and debugging. Ah, I had a different kernel make process in mind.
    Oracle doesn't ship
    debuginfo package. Of course, this is when I build a
    "complete suite" of kernel rpm packages using
    unmodified spec file. And, to be honest, it takes
    much more than an hour, maybe even two hours. Another
    thing is compiling single i686 kernel without
    building a package. But it also takes at least half
    an hour. Anyway the time significantly depends on how
    many modules are selected to be built in.That what I was looking for.
    What's your time to build a single kernel (which
    version?) with default set of modules ? On which
    hardware ? I've only access to a root server right now, which is
    cat /proc/cpuinfo | grep "model name"
    model name      : AMD Athlon(tm) 64 Processor 3700+with about 2GB of RAM
    free -m
                 total       used       free     shared    buffers     cached
    Mem:          2024       1957         67          0        368       1291
    -/+ buffers/cache:        297       1727
    Swap:         3827         24       3803under
    uname -a
    Linux base 2.6.22-gentoo-r5 #5 PREEMPT Mon Sep 10 22:32:37 CEST 2007 i686 AMD Athlon(tm) 64 Processor 3700+ AuthenticAMD GNU/LinuxThis is what i did
    cd /usr/src/linux
    make clean
    time nice -n  19 genkernel --lvm2 --makeopts="-j2" --oldconfig all
    * Running with options: --lvm2 --makeopts=-j2 --oldconfig all
    * Linux Kernel 2.6.22-gentoo-r5 for x86...*
    mount: /boot mounted successfully!
    * config: >> Running oldconfig...
    * config: --no-clean is enabled; leaving the .config alone.
    *         >> Compiling 2.6.22-gentoo-r5 bzImage...
    *         >> Compiling 2.6.22-gentoo-r5 modules......
    real    17m30.582s
    user    16m8.480s
    sys     1m9.000sWhat could helped here was that I've switched off some modules and (maybe) the use of ccache.
    C.

  • Execute more than one times external command of OS Windows NT

    Hi all,
            I have to execute more than one times from an ABAP program an exernal command of Operating system windows NT . In order to reach this goal I found the standard program RSBDCOS0 and it works very well if you execute it one time.  Unfortunately I have to execute this external command more than one times. Is there any soluiton?
    thank to all
                        Gino Bonfiglioli

    Well why are you using this report RSBDCOS0 to execute external commands?
    Because SAP provides certain function modules that can be used to call operating system commands; so you can use these functions to do your task.
    Just check the following link
    http://help.sap.com/saphelp_nw2004s/helpdata/en/fa/0971e1543b11d1898e0000e8322d00/frameset.htm
    As you mentioned that you need to execute multiple times, you can easily put the call to function module inside a loop or do/enddo structure. for example consider the following
    do 3 times.
    call function 'SXPG_COMMAND_EXECUTE'                           
          exporting                                                 
               commandname                   = 'Put operating system command here'          
               additional_parameters         = full_filename        
               targetsystem                  = targtsys  
               stdout                        = 'X'                  
               stderr                        = 'X'                  
               terminationwait               = 'X'                  
          importing                                                 
               status                        = sxpg_status          
               exitcode                      = sxpg_exitcode        
          tables                                                    
               exec_protocol                 = sxpg_results         
          exceptions                                                
               no_permission                 = 1                    
               command_not_found             = 2                    
               parameters_too_long           = 3                    
               security_risk                 = 4                    
               wrong_check_call_interface    = 5                    
               program_start_error           = 6                    
               program_termination_error     = 7                    
               x_error                       = 8                    
               parameter_expected            = 9                    
               too_many_parameters           = 10                   
               illegal_command               = 11                   
               wrong_asynchronous_parameters = 12                   
               cant_enq_tbtco_entry          = 13                   
               jobcount_generation_error     = 14                   
               others                        = 15.                  
    write the results
    loop at sxpg_results.
    write:/ sxpg_results-length,
             sxpg_results-message.
    endloop.
    refresh sxpg_results.
    enddo.

  • Is an Intel iMac faster than PowerMac G4 1.25 Ghz single proccesor?

    Hi, I'm owner of a PowerMac G4 1.25 Ghz single proccesor, working whit Adobe Creative Suite 1. I'm wonderin if the iMac Intel Core Duo is faster than my PowerMac G4? Thanks for your support.

    I have a 1.25 G4 Mac Mini and the new intel Mac. The answer to your question is that the iMac will be loads faster for anything that is a native Universal Binary like iLife programs. It will also be faster for games that are not the latest thing, but that are accelerated by its fairly beefy 3d card - so faster for Halo, Call of Duty - Unreal Tournamaent1 etc. (other Unreal tournaments not good under Rosetta - but 2004 now has a UB so that should whizz when you download the patch, if you are into such things).
    However it will be slower for power applications running under Rosetta - ie: Photoshop or Dreamweaver. Until the Universal Binary versions come - then it will be much faster for those too.

  • Don't want Call Waiting or Voice Mail

    The only features I need with my digital phone service are unlimited calling and Caller ID.  I've asked to have Voice Mail disabled because I use a traditional answering machine connected to my wireless phones.  I've asked to have Call Waiting disabled because giving the caller a ring tone when my phone is busy confuses them, and the messages are often misleading.  Right now, service says they have disabled these features, yet if my phone is called while it is in use the caller hears the ring tone ten times then gets a message that says "The number you are calling is not answering.  Please call again later."  All I want is for anyone calling me when I'm on the phone to receive the time-honored "Busy" signal.  Can anyone reassure me that their phone has these two features turned off and works this way, to give me hope that Comcast can fix the problem? 

    Just to close the book on this:  Apparently even Comcast has trouble turning off Call Waiting and Mailbox services.  I was told by three seperate level 1 service techs that my problem was caused by my account showing those services had not been shut off, and that they had finally fixed it.  After the fourth attempt by a level 2 tech, the problem was referred to "level 3 support", so exalted that you can't even talk to them, just wait (for up to 72 hours) for them to fix it and let you know.  Never heard a word from them, but tested it five days later and problem was fixed.  If I call myself with a cell while I'm on the phone, I immediately hear that beautiful busy tone on the cell, no confusing voice mail messages.  And also there are no "call waiting" beeps on main phone.  If I call myself with a cell while phone is not in use, my traditional answering machine picks up on fourth ring and takes a message, just like it should.     However, you can't make the test by just calling yourself on the main phone.  Comcast treats that differently.  Phone rings once, then a voice says "The mailbox number that you are trying to reach is not in service.  Goodbye."  I can live with that because it will never happen.  But Comcast is ruining an easy way for customers to check whether these features have been disabled. If Comcast were really interested in improving the "call waiting" feature they would detect that the destination phone was in use, ring a few times (no more than six) to give the called party a chance to switch over, then give the calling party a message that says "The number you are trying to reach is busy.  Please try again later. Goodbye."  

  • Why is Mac Pro 2.66 only 1.3x faster than 2.7 G5 on CPU intensive stuff?

    I produce DVDs so my Compressor DVCam -> MPEG2 encoding is the most time consuming task. Take the MacWorld benchmarks, I was dissappointed the QC 2.66 was a third faster than a DC 2.7 G5 running Compressor.
    I would have expected almost 2x as fast, basically halving encoding times. The Mac Pro took 107s vs G5 137s only 1.28x as fast OR put another way jobs complete in 78% of the time taken for the G5.
    This is key reason for me to have just sold a G5 DC 2.3...but I'm dissappointed with these early indicators. Would it be reasonable to assume Apple have not optimised Compressor for Intel - surely not at this late stage?
    G4 Dual Gigabit   Mac OS X (10.4.7)   ATI 9800 Pro

    Terpstar,
    I was wondering if you have had a chance to use Motion yet. I have a MBP, and using Zapfino fonts with SciFi Glow crashes my system every time. I would be interested to see if this is the case on other intel based systems. This has led to a failure of my main logic board twice over the last month. See my thread:
    http://discussions.apple.com/thread.jspa?threadID=614641&tstart=25
    Also, of the two GB ram I have installed, FCP doesn't seem to utilize more than 100MB of RAM. Although the VM size is several GB for the app. I noticed that in order to utilize both cores on my MBP, Airport had to be turned off.
    Also, as Ned Snowing was saying, there is no doubt that there are going to be many software bugs that must be sorted out. Especially since this program is being adapted for intel macs, and not re-written.

  • Execute more than one thing at the same time.

    Hey why every time i want to execute more than one thing(create tables) at the same time
    it say "ORA-00911: invalid character" , i dont wanna execute one thing at a time. check the pic please
    http://img370.imageshack.us/img370/5120/oraclesw5.jpg
    but in my school i use something called isqlplus and i can execute all the things i want at the same time with-out any error.
    im using right-know
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Product
    PL/SQL Release 10.2.0.1.0 - Production
    CORE 10.2.0.1.0 Production
    TNS for 32-bit Windows: Version 10.2.0.1.0 - Production
    NLSRTL Version 10.2.0.1.0 - Production

    1) Unless you are using something like DBMS_JOB, SQL*Plus, iSQL*Plus, the Oracle XE web-based front-end, or any other application is going to execute one SQL statement at a time.
    2) There is a dedicated XE forum-- the folks over there might know how to run a script using this particular front-end.
    3) Assuming the XE database is on your local system, you should be able to log in via SQL*Plus and use that utility to run your scripts.
    Justin

Maybe you are looking for