Writing to disk with Oracle...

Hi,
does Oracle use the operating system input/output system calls in order to write/read from/to disk, or does it use its own routines. Does it use the caching mechanisms of UNIX-like systems? Since Oracle has its own caching mechanisms for writing asynchronously to the datafiles, is Oracle using the operating system specific caches?
Thanks in advance

If you are creating Oracle data files on a standard Linux file system, Oracle will use the operating system commands to read and write data. Oracle also has the ability to use "raw" disks, where the operating system does not have a file system, using its own I/O routines. Raw disks are historically slightly faster, but more effort to manage. Oracle can also use ASM in 10g, which adds to the options and the complexity.
I believe you can enable asynchronous I/O on Linux, but I'm not expert enough to really understand how that interacts with the OS caching algorithms.
Justin
Distributed Database Consulting, Inc.
http://www.ddbcinc.com/askDDBC

Similar Messages

  • Writing request XQUERY with oracle 11G

    Hi,
    What is exactly the best way to make an XMLTYPE object (column with type XMLType) request with Oracle 11 g ? (Oracle 11g  Release 11.2.0.3.0)
    In different books, we see a lot of different requests.
    Simple example :
    create table contact_file_tbl(filename varchar2(255), xml xmltype);
    insert into contact_file_tbl (filename, xml)
    values ('john_smith.xml', xmltype(bfilename('XML_DIR', 'john_smith.xml'),nls_charset_id('AL32UTF8')));
    Ok for now
    But impossible to make a good query :
    select XMLQuery('declare namespace book="http://xmlbook.com/sample/contact.xsd"; (:
    for $c in /book:contact
    return $c/book:first_name' PASSING a.xml RETURNING CONTENT)
    from contact_file_tbl a;
    returns nothing.
    why this request works ?
    SELECT filename, XMLQuery('toto'
             PASSING xml RETURNING CONTENT) dataResult
      FROM contact_file_tbl;
    Could you send me a request with this example.
    Here the XML document :
    <?xml version="1.0" encoding="UTF-8"?>
    <contact xmlns="http://xmlbook.com/sample/contact.xsd" id="1">
       <category>customer</category>
       <first_name Chinese="约翰">John</first_name>
       <last_name>Smith</last_name>
       <email>[email protected]</email>
       <phone>(512)781-9230</phone>
       <cellphone>(512)781-9230</cellphone>
       <address>
           <street1>1234 sunflower road</street1>
           <city>austin</city>
           <state>texas</state>
           <zipcode>78701</zipcode>
           <country>USA</country>
       </address>
       <references>
         <reference relationship="account manager, tech sales">
           <first_name>robert</first_name>
           <last_name>tian</last_name>
             <email>[email protected]</email>
         </reference>
         <reference relationship="account manager, app sales">
           <first_name>Richard</first_name>
           <last_name>Liu</last_name>
           <email>[email protected]</email>
         </reference>
       </references>
    </contact>
    Thank you very much

    This works fine for me :
    SQL> select XMLQuery(
      2  'declare namespace book="http://xmlbook.com/sample/contact.xsd"; (::)
      3  for $c in /book:contact
      4  return $c/book:first_name'
      5  PASSING a.xml
      6  RETURNING CONTENT
      7  )
      8  from contact_file_tbl a;
    XMLQUERY('DECLARENAMESPACEBOOK
    <first_name xmlns="http://xmlbook.com/sample/contact.xsd" Chinese="??">John</fir
    SQL>
    SQL> select XMLCast(
      2  XMLQuery(
      3  'declare namespace book="http://xmlbook.com/sample/contact.xsd"; (::)
      4  for $c in /book:contact
      5  return $c/book:first_name'
      6  PASSING a.xml
      7  RETURNING CONTENT
      8  )
      9  as varchar2(30)
    10  ) as first_name
    11  from contact_file_tbl a;
    FIRST_NAME
    John
    What's your client tool/version?

  • On Sun fire v490 - Solaris 10 with Oracle 8.1.7.4 & Sybase 12.0

    Hi,
    We are going to upgrade our server with this configuration -
    Sun Fire V490     2 x 1.05 GHz UltraSPARC IV CPU
    8096MB RAM     2 x73GB local disk
    2x FC 2GB Sun/QLogic HBAs
    DAT72
    On one machine we will have Sun Solaris v10 with
    Oracle DB v8.1.7.4 & Second one will be Sun Solaris v10 with Sybase DB v12.0.0.6.
    Now our question is - Sun fire have Hyper-thread CPUs ��� will the O/S and databases (Oracle and Sybase) view the proposed system as a true 4 CPU platform? Will parameters used to tune the database such as Sybase max online engines still operate in the same manner as before?
    Our old machine configuration was - Sun E450     4x400MHz CPU     1024MB RAM     2 x18; 8x36GB disks

    Questions on Oracle and Sybase should be directed to a database forum, this forum is for Sun hardware support.
    Here is a link to a DB forum I look at from time to time:
    http://www.dbforums.com/index.php
    The topic of tuning Oracle or Solaris is way beyond the scope of this forum, I have attempted to go into it before but didn't get any feedback and I would only like to spend lots of time on it if I was being paid!!! On the memory side, keep in mind that Oracle 9i 64-bit can address a maximum of 2 ^ 64 ( 16777216 TB ) memory, prior to that the DBA had to define memory parameters in init.ora. To be honest the last time I worked with a Oracle 8 database I shut a HP K class server down permanently that had been migrated to Oracle 9i on Solaris by an Oracle consultant and I can't remember all the tuning trick etc.

  • Cannot install DBD::Oracle perl module with Oracle 11.2.0.3 client

    Hi
    I have been trying to install DBD::Oracle perl module on a Centos 5.7 server which is running Oracle client version 11.2.0.3. I have installed all dependencies prior to attempting installing the same and I am running into the following error -
    /usr/bin/ld: skipping incompatible /app/utiba/oracle/client/lib//libclntsh.so when searching for -lclntsh
    /usr/bin/ld: cannot find -lclntsh
    collect2: ld returned 1 exit status
    make: *** [blib/arch/auto/DBD/Oracle/Oracle.so] Error 1Here are the complete details of my attempt of installation of the perl module -
    [root@tch1testuview1 DBD-Oracle-1.50]# perl Makefile.PL
    Using DBI 1.52 (for perl 5.008008 on x86_64-linux-thread-multi) installed in /usr/lib64/perl5/site_perl/5.8.8/x86_64-linux-thread-multi/auto/DBI/
    Configuring DBD::Oracle for perl 5.008008 on linux (x86_64-linux-thread-multi)
    Remember to actually *READ* the README file! Especially if you have any problems.
    Installing on a linux, Ver#2.6
    Using Oracle in /app/utiba/oracle/client
    DEFINE _SQLPLUS_RELEASE = "1102000300" (CHAR)
    Oracle version 11.2.0.3 (11.2)
    Found /app/utiba/oracle/client/rdbms/lib/ins_rdbms.mk
    Using /app/utiba/oracle/client/rdbms/lib/ins_rdbms.mk
    Your LD_LIBRARY_PATH env var is set to '/app/utiba/oracle/client/lib:/lib:/usr/lib'
    Reading /app/utiba/oracle/client/rdbms/lib/ins_rdbms.mk
    Reading /app/utiba/oracle/client/rdbms/lib/env_rdbms.mk
    WARNING: Oracle /app/utiba/oracle/client/rdbms/lib/ins_rdbms.mk doesn't define a 'build' rule.
    WARNING: I will now try to guess how to build and link DBD::Oracle for you.
             This kind of guess work is very error prone and Oracle-version sensitive.
             It is possible that it won't be supported in future versions of DBD::Oracle.
             *PLEASE* notify dbi-users about exactly _why_ you had to build it this way.
    Found header files in /app/utiba/oracle/client/rdbms/public.
    client_version=11.2
    DEFINE= -Wall -Wno-comment -DUTF8_SUPPORT -DORA_OCI_VERSION=\"11.2.0.3\" -DORA_OCI_102 -DORA_OCI_112
    Checking for functioning wait.ph
    System: perl5.008008 linux builder10.centos.org 2.6.18-53.el5 #1 smp mon nov 12 02:14:55 est 2007 x86_64 x86_64 x86_64 gnulinux
    Compiler:   gcc -O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic -D_REENTRANT -D_GNU_SOURCE -fno-strict-aliasing -pipe -Wdeclaration-after-statement -I/usr/local/include -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -I/usr/include/gdbm
    Linker:     /usr/bin/ld
    Sysliblist: -ldl -lm -lpthread -lnsl -lirc -lipgo
    Oracle makefiles would have used these definitions but we override them:
      CC:       $(GCC)
      CFLAGS:   $(GFLAG) $(OPTIMIZE) $(CDEBUG) $(CCFLAGS) $(PFLAGS)\
            $(SHARED_CFLAG) $(USRFLAGS)
               [$(GFLAG) -O2 $(CDEBUG) $(CCFLAGS) -I/app/utiba/oracle/client/rdbms/demo -I/app/utiba/oracle/client/rdbms/public -I/app/utiba/oracle/client/plsql/public -I/app/utiba/oracle/client/network/public -DLINUX -D_GNU_SOURCE -D_LARGEFILE64_SOURCE=1 -D_LARGEFILE_SOURCE=1 -DSLTS_ENABLE -DSLMXMX_ENABLE -D_REENTRANT -DNS_THREADS -DLDAP_CM $(LPFLAGS) $(PLSQLNCGFLAGS) $(USRFLAGS)]
      LDFLAGS:  -o $@ $(LDOBJSZ) $(LDPATHFLAG)$(PRODLIBHOME) $(LDPATHFLAG)$(LIBHOME) $(LDPATHFLAG)$(LIBHOME)stubs/ $(LDPATHFLAG)$(LIBHOME) $(LLIBIRC)
               [-o $@ -m32 -L/app/utiba/oracle/client/rdbms/lib/ -L$(LIBHOME) -L$(LIBHOME)stubs/ -L$(LIBHOME) -lirc -lipgo ]
    Linking with  -lclntsh -ldl -lm -lpthread -lnsl -lirc -lipgo -ldl -lm  [from $(OCISHAREDLIBS)]
    LD_RUN_PATH=/app/utiba/oracle/client/lib
    Using DBD::Oracle 1.50.
    Using DBD::Oracle 1.50.
    Using DBI 1.52 (for perl 5.008008 on x86_64-linux-thread-multi) installed in /usr/lib64/perl5/site_perl/5.8.8/x86_64-linux-thread-multi/auto/DBI/
    Writing Makefile for DBD::Oracle
    ***  If you have problems...
         read all the log printed above, and the README and README.help.txt files.
         (Of course, you have read README by now anyway, haven't you?)
    [root@tch1testuview1 DBD-Oracle-1.50]# make
    Skip blib/lib/DBD/Oracle/Troubleshooting/Cygwin.pod (unchanged)
    Skip blib/lib/DBD/Oracle.pm (unchanged)
    Skip blib/lib/oraperl.ph (unchanged)
    Skip blib/arch/auto/DBD/Oracle/Oracle.h (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Vms.pod (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Linux.pod (unchanged)
    Skip blib/lib/DBD/Oracle/GetInfo.pm (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting.pod (unchanged)
    Skip blib/arch/auto/DBD/Oracle/dbdimp.h (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Hpux.pm (unchanged)
    Skip blib/arch/auto/DBD/Oracle/ocitrace.h (unchanged)
    Skip blib/lib/Oraperl.pm (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Sun.pod (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Macos.pod (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Aix.pod (unchanged)
    Skip blib/lib/DBD/Oracle/Object.pm (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Win64.pod (unchanged)
    Skip blib/arch/auto/DBD/Oracle/mk.pm (unchanged)
    Skip blib/lib/DBD/Oracle/Troubleshooting/Win32.pod (unchanged)
    gcc -c  -I/app/utiba/oracle/client/rdbms/public -I/app/utiba/oracle/client/rdbms/demo -I/app/utiba/oracle/client/rdbms/public -I/app/utiba/oracle/client/plsql/public -I/app/utiba/oracle/client/network/public -I/usr/lib64/perl5/site_perl/5.8.8/x86_64-linux-thread-multi/auto/DBI -D_REENTRANT -D_GNU_SOURCE -fno-strict-aliasing -pipe -Wdeclaration-after-statement -I/usr/local/include -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -I/usr/include/gdbm -O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic   -DVERSION=\"1.50\" -DXS_VERSION=\"1.50\" -fPIC "-I/usr/lib64/perl5/5.8.8/x86_64-linux-thread-multi/CORE"  -Wall -Wno-comment -DUTF8_SUPPORT -DORA_OCI_VERSION=\"11.2.0.3\" -DORA_OCI_102 -DORA_OCI_112 Oracle.c
    gcc -c  -I/app/utiba/oracle/client/rdbms/public -I/app/utiba/oracle/client/rdbms/demo -I/app/utiba/oracle/client/rdbms/public -I/app/utiba/oracle/client/plsql/public -I/app/utiba/oracle/client/network/public -I/usr/lib64/perl5/site_perl/5.8.8/x86_64-linux-thread-multi/auto/DBI -D_REENTRANT -D_GNU_SOURCE -fno-strict-aliasing -pipe -Wdeclaration-after-statement -I/usr/local/include -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -I/usr/include/gdbm -O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic   -DVERSION=\"1.50\" -DXS_VERSION=\"1.50\" -fPIC "-I/usr/lib64/perl5/5.8.8/x86_64-linux-thread-multi/CORE"  -Wall -Wno-comment -DUTF8_SUPPORT -DORA_OCI_VERSION=\"11.2.0.3\" -DORA_OCI_102 -DORA_OCI_112 dbdimp.c
    dbdimp.c: In function ‘fb_ary_cb_alloc’:
    dbdimp.c:324: warning: comparison is always false due to limited range of data type
    dbdimp.c:325: warning: comparison is always false due to limited range of data type
    dbdimp.c:326: warning: comparison is always false due to limited range of data type
    dbdimp.c:327: warning: comparison is always false due to limited range of data type
    dbdimp.c:328: warning: comparison is always false due to limited range of data type
    dbdimp.c: In function ‘fb_ary_alloc’:
    dbdimp.c:344: warning: comparison is always false due to limited range of data type
    dbdimp.c:345: warning: comparison is always false due to limited range of data type
    dbdimp.c:346: warning: comparison is always false due to limited range of data type
    dbdimp.c:347: warning: comparison is always false due to limited range of data type
    gcc -c  -I/app/utiba/oracle/client/rdbms/public -I/app/utiba/oracle/client/rdbms/demo -I/app/utiba/oracle/client/rdbms/public -I/app/utiba/oracle/client/plsql/public -I/app/utiba/oracle/client/network/public -I/usr/lib64/perl5/site_perl/5.8.8/x86_64-linux-thread-multi/auto/DBI -D_REENTRANT -D_GNU_SOURCE -fno-strict-aliasing -pipe -Wdeclaration-after-statement -I/usr/local/include -D_LARGEFILE_SOURCE -D_FILE_OFFSET_BITS=64 -I/usr/include/gdbm -O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic   -DVERSION=\"1.50\" -DXS_VERSION=\"1.50\" -fPIC "-I/usr/lib64/perl5/5.8.8/x86_64-linux-thread-multi/CORE"  -Wall -Wno-comment -DUTF8_SUPPORT -DORA_OCI_VERSION=\"11.2.0.3\" -DORA_OCI_102 -DORA_OCI_112 oci8.c
    oci8.c: In function ‘ora_blob_read_mb_piece’:
    oci8.c:1804: warning: comparison is always false due to limited range of data type
    oci8.c: In function ‘describe_obj_by_tdo’:
    oci8.c:3256: warning: comparison is always false due to limited range of data type
    oci8.c: In function ‘ora_describe’:
    oci8.c:3480: warning: comparison is always false due to limited range of data type
    oci8.c: In function ‘ora_st_fetch’:
    oci8.c:3953: warning: unused variable ‘imp_xxh’
    Running Mkbootstrap for DBD::Oracle ()
    chmod 644 Oracle.bs
    rm -f blib/arch/auto/DBD/Oracle/Oracle.so
    LD_RUN_PATH="/app/utiba/oracle/client/lib" gcc  -shared -O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic Oracle.o dbdimp.o oci8.o  -o blib/arch/auto/DBD/Oracle/Oracle.so  \
               -L/app/utiba/oracle/client/lib/ -lclntsh -ldl -lm -lpthread -lnsl -lirc -lipgo -ldl -lm      \
    /usr/bin/ld: skipping incompatible /app/utiba/oracle/client/lib//libclntsh.so when searching for -lclntsh
    /usr/bin/ld: cannot find -lclntsh
    collect2: ld returned 1 exit status
    make: *** [blib/arch/auto/DBD/Oracle/Oracle.so] Error 1I am sure that installing a previous(older) version of oracle client will fix this issue i.e. either 11.2.0.1 or older (that is tested) but I want to know if it is possible to resolve the above issue with a work around and not having to install a previous version.
    Fyi.. I've tried to install DBD::Oracle version 1.24 and 1.50 but that doesn't seem to make any difference. I couldn't find any related thread either for this error so I desperately need a solution to the same(if any).
    If anyone can help here, would really appreciate it or else, I will have no choice but to install previous version of oracle client.
    Thanks
    Edited by: 956308 on Aug 31, 2012 6:55 AM
    Edited by: 956308 on Aug 31, 2012 6:57 AM

    Contents of the response file:
    oracle.install.responseFileVersion=http://www.oracle.com/2007/install/rspfmt_clientinstall_response_schema_v11_2_0
    ORACLE_HOSTNAME=slc00caz
    UNIX_GROUP_NAME=oracle
    INVENTORY_LOCATION=/home/psadm1/oraInventory
    SELECTED_LANGUAGES=en
    ORACLE_HOME=/opt/oracle/psft/pt/oracle-client/11.2.0.x
    ORACLE_BASE=/opt/oracle/psft/pt/oracle-client
    oracle.install.client.installType=Custom
    oracle.install.client.upgrading=false
    oracle.install.client.customComponents=oracle.rdbms.util:11.2.0.3.0,oracle.javavm.client:11.2.0.3.0,oracle.sqlplus:11.2.0.3.0,oracle.dbjava.jdbc:11.2.0.3.0,oracle.rdbms.oci:11.2.0.3.0,oracle.xdk:11.2.0.3.0,oracle.network.aso:11.2.0.3.0,oracle.assistants.oemlt:11.2.0.3.0,oracle.network.client:11.2.0.3.0,oracle.odbc:11.2.0.3.0,oracle.dbdev:11.2.0.3.0
    oracle.install.client.schedulerAgentHostName=
    oracle.install.client.schedulerAgentPortNumber=
    oracle.installer.autoupdates.option=SKIP_UPDATES
    oracle.installer.autoupdates.downloadUpdatesLoc=
    AUTOUPDATES_MYORACLESUPPORT_USERNAME=
    AUTOUPDATES_MYORACLESUPPORT_PASSWORD=
    PROXY_HOST=
    PROXY_PORT=0
    PROXY_USER=
    PROXY_PWD=
    PROXY_REALM=

  • Help, I installed parallel and windows, and opened videos (on an external hard disk) with windows, and now cannot open them with Mac, and use i movies.  How do I reverse this?

    Help, I installed parallel and windows, and opened videos (on an external hard disk) with windows, and now cannot open them with Mac, and use i movies.  How do I reverse this?

    Paragraphs help the reader, just saying.
    What are you actually trying to do? Why do you need "DVD size" folders if you're creatng an Archive on an external hard drive?
    Also, I understand burning folders instead of individual photos or albums gives you more space on a DVD.
    It makes no difference at all.
    How can I keep the captions on my photos (not events) when I print or transfer to a new folder?
    You can export while writing the metadata to the file in most cases.
    Tne only way I know is to copy photos from the Windows to folders on my HD, then copy and paste photos or albums or events to HD, then create a new folder and fill it to the appropriate DVD size. 
    HD? Another HD? You use HD there in a way that makes no sense. What's the point of 'DVD size'?
    I also thought I should download future photos to my Canon programming that comes with EOS cameras and edit and then decide which photos to save to iPhoto library and which to save only to my archival HD?
    Does this have any relation to the previous questions? You know you can delete from iPhoto, right?
    You're somewhat confused and you need to
    1. Decide what exactly you want to do
    2. Then go about it.
    If I understand you correctly - and there's no guarantee that I do -  the easiest thing is to bring all your Photos into iPhoto and do everything from there.
    You can have an archive on an external disk, you can sort and select and edit with or via iPhoto
    Regards
    TD

  • Poor performance with Oracle Spatial when spatial query invoked remotely

    Is anyone aware of any problems with Oracle Spatial (10.2.0.4 with patches 6989483 and 7003151 on Red Hat Linux 4) which might explain why a spatial query (SDO_WITHIN_DISTANCE) would perform 20 times worse when it was invoked remotely from another computer (using SQLplus) vs. invoking the very same query from the database server itself (also using SQLplus)?
    Does Oracle Spatial have any known problems with servers which use SAN disk storage? That is the primary difference between a server in which I see this poor performance and another server where the performance is fine.
    Thank you in advance for any thoughts you might share.

    OK, that's clearer.
    Are you sure it is the SQL inside the procedure that is causing the problem? To check, try extracting the SQL from inside the procedure and run it in SQLPLUS with
    set autotrace on
    set timing on
    SELECT ....If the plans and performance are the same then it may be something inside the procedure itself.
    Have you profiled the procedure? Here is an example of how to do it:
    Prompt Firstly, create PL/SQL profiler table
    @$ORACLE_HOME/rdbms/admin/proftab.sql
    Prompt Secondly, use the profiler to gather stats on execution characteristics
    DECLARE
      l_run_num PLS_INTEGER := 1;
      l_max_num PLS_INTEGER := 1;
      v_geom    mdsys.sdo_geometry := mdsys.sdo_geometry(2002,null,null,sdo_elem_info_array(1,2,1),sdo_ordinate_array(0,0,45,45,90,0,135,45,180,0,180,-45,45,-45,0,0));
    BEGIN
      dbms_output.put_line('Start Profiler Result = ' || DBMS_PROFILER.START_PROFILER(run_comment => 'PARALLEL PROFILE'));  -- The comment name can be anything: here it is related to the Parallel procedure I am testing.
      v_geom := Parallel(v_geom,10,0.05,1);  -- Put your procedure call here
      dbms_output.put_line('Stop Profiler Result = ' || DBMS_PROFILER.STOP_PROFILER );
    END;
    SHOW ERRORS
    Prompt Finally, report activity
    COLUMN runid FORMAT 99999
    COLUMN run_comment FORMAT A40
    SELECT runid || ',' || run_date || ',' || run_comment || ',' || run_total_time
      FROM plsql_profiler_runs
      ORDER BY runid;
    COLUMN runid       FORMAT 99999
    COLUMN unit_number FORMAT 99999
    COLUMN unit_type   FORMAT A20
    COLUMN unit_owner  FORMAT A20
    COLUMN text        FORMAT A100
    compute sum label 'Total_Time' of total_time on runid
    break on runid skip 1
    set linesize 200
    SELECT u.runid || ',' ||
           u.unit_name,
           d.line#,
           d.total_occur,
           d.total_time,
           text
    FROM   plsql_profiler_units u
           JOIN plsql_profiler_data d ON u.runid = d.runid
                                         AND
                                         u.unit_number = d.unit_number
           JOIN all_source als ON ( als.owner = 'CODESYS'
                                   AND als.type = u.unit_type
                                   AND als.name = u.unit_name
                                AND als.line = d.line# )
    WHERE  u.runid = (SELECT max(runid) FROM plsql_profiler_runs)
    ORDER BY d.total_time desc;Run the profiler in both environments and see if you can see where the slowdown exists.
    regards
    Simon

  • NotSerializableException when writing to disk

    Hi,
    I am having a problem with simply writing an object to disk. Have a class called DataHolder which implements serializable. This class consists of about a dozen strings and nothing else:
    class DataHolder implements Serializable{
    String a="abc";
    String b="def";
    etc..........
    I add instances to an ArrayList then pull them out.
    DataHolder holder=(DataHolder)list.get(i);
    When I try to write them to disk with:
    ObjectOutputStream objectOut = new ObjectOutputStream(new FileOutputStream("myFile"));
    objectOut.writeObject(holder);
    I get an error:
    java.io.NotSerializableException: DataHolder
    I know there must be a simple explanation for this but I don't see it. I think DataHolder should be serializable since it implements Serializable, right?
    thanks
    john

    Hi john,
    The error[b] java.io.NotSerializableException:
    thrown when an instance is required to have a Serializable interface. The serialization runtime or the class of the instance can throw this exception.
    For this the argument should be the name of the class.
    Try by doing
    class DataHolder implements java.io.Serializable {
    String a="abc";
    String b="def";
    Hope this will help you.
    Regards,
    Anil.
    Tehnical Support Engineer.

  • Why is Premiere 6 giving a "Error writing file (disk full?)" message?

    Hello helpful experts,
    I'm using Premiere 6 on a Windows 7 machine.
    I have a file imported in to the timeline and it is set to play in reverse via the speed setting being set to -100 and it plays back in the preview window correctly, in reverse, when I click the play button.
    But when I do a  "File>Export Timeline>Movie" command I get an error message saying "An error occurred while making the movie. Error writing file (disk full?)"... but my hard drive is not nearly full, with over 30 gigs left open on it... and my RAM usage gadget shows over half of my RAM capacity is unused.
    Now previously, weeks ago on a different project, I was getting this same error message for a while when I would hit enter to preview my movie on the timeline (but that worked itself out somehow and eventually stopped giving me that error message).  But when I went to save and export, then I would not get the error message... now, it is just the opposite... I'm now getting the error message when trying to do an export and save the movie.
    And also now when I hit enter to preview, as opposed to just clicking the play button on the preview window, Premiere is going through the process of saving the movie since it is saving a temporary preview file... hopefully, since I'm getting this error message, I can find the temporary preview file and save it as my exported AVI file.
    But I still shouldn't be getting this error message when trying to save, or when trying to preview, as I did for a while a couple of weeks ago.
    Does anyone know about the erroneous error message that says, "An error occurred while making the movie. Error writing file (disk full?)"
    Thanks,
    numetro

    Depending on your hard drive size, P6 may be confused... since drives were not as large when it was written
    For Win7 64bit Pro or Ultimate (or Enterprise?) this is ONE example of using Virtual XP http://forums.adobe.com/thread/702693?tstart=0
    This is saved from an OLD message that may help
    File size limits by disk format type
    Fat = 2 Gig maximum file size
    Fat32 = 4 Gig
    NTFS = ?Terrabytes?
    Message saved from Premiere forum
    John Acocks - 07:52am Apr 17, 2003 Pacific
    I am unable to complete captures because I get a message
    that says "capture stopped because disk was full." What
    does this mean?
    Daniel Conklin - 08:02am Apr 17, 2003 Pacific (#1 of 2)   
    If you are using a Pinnacle card, a possible solution to
    this is to right click on the recycle bin, and set the
    percentage on your media drive to less than 2%.
    There a several other possible causes as well, but you
    need to supply some info about your system. How are you
    capturing and what format is your video? How big is
    your hard drive?
    Jeff Schell - 09:57am Apr 17, 2003 Pacific (#2 of 2)   
    this is one of my favorite error messages. Basically,
    you've chosen an incorrect setting in your DV settings.
    (Odds are you have DV settings). If so, read on:
    1. The DV format is very rigid. Meaning that if you
       make changes to the format, it cannot export as
       requested, because your movie file must match every
       single spec exactly.
    2. When Premiere goes to export a project, the DV codec
       says "Whoa. I can't do anything, because the settings
       don't match."
    3. Then Premiere comes back to you and has to report an
       error. (Don't shoot the messenger). Unfortunately,
       Premiere doesn't know why it wasn't able to export a
       movie. So it says to you, "Um... I can't make the movie.
       Um... maybe your hard drives are full?"
    It's a complete guess. And has nothing to do with your harddrives.
    Check your project settings and/or your export settings. Make
    sure you have chosen one of the "presets" that will specify
    every correct setting for you. Make sure it is one of the DV
    presets, if you are using DV video, or make sure it is one of
    the presets supplied by your capture card manufacturer (in this
    case I'm going to guess Pinnacle or Matrox.)
    To choose a preset, go to Project > Settings Viewer, and then
    click on the project settings column, and then click 'load' to
    load a correct preset. Do the same with export settings.

  • IMP-00017: following statement failed with ORACLE error 1658

    Hi All,
    Please help me to find the reason of this error while importing the export file using parameters full=Y,ignore=n.
    IMP-00003: ORACLE error 1658 encountered
    ORA-01658: unable to create INITIAL extent for segment in tablespace USERS
    IMP-00017: following statement failed with ORACLE error 1658:
    "CREATE TABLE "TABLE_NAME" ("SEGMENTID" VARCHAR2(10) NOT NULL EN"
    "ABLE, "DRIVERID" VARCHAR2(10) NOT NULL ENABLE, "VEHICLEID" VARCHAR2(10) NOT"
    " NULL ENABLE, "START_TIME" VARCHAR2(6) NOT NULL ENABLE, "END_TIME" VARCHAR2"
    "(6) NOT NULL ENABLE, "REPORT_TIME" VARCHAR2(6) NOT NULL ENABLE, "DISPOSITIO"
    "N" VARCHAR2(1), "S_LOCATION1" VARCHAR2(12), "S_LOCATION2" VARCHAR2(60) NOT "
    "NULL ENABLE, "S_LOCATION3" VARCHAR2(12), "S_LOCATION_ENT" VARCHAR2(30), "S_"
    "CITYTOWN" VARCHAR2(30) NOT NULL ENABLE, "S_ZIP" VARCHAR2(10) NOT NULL ENABL"
    "E, "S_STATEPRO" VARCHAR2(4) NOT NULL ENABLE, "S_COUNTRY" VARCHAR2(2), "S_CO"
    "MNTS" VARCHAR2(1), "S_GRID_LAT" NUMBER(10, 6), "S_GRID_LONG" NUMBER(10, 6),"
    " "E_LOCATION1" VARCHAR2(12), "E_LOCATION2" VARCHAR2(60) NOT NULL ENABLE, "E"
    "_LOCATION3" VARCHAR2(12), "E_LOCATION_ENT" VARCHAR2(30), "E_CITYTOWN" VARCH"
    "AR2(30) NOT NULL ENABLE, "E_ZIP" VARCHAR2(10) NOT NULL ENABLE, "E_STATEPRO""
    " VARCHAR2(4) NOT NULL ENABLE, "E_COUNTRY" VARCHAR2(2), "E_COMNTS" VARCHAR2("
    "1), "E_GRID_LAT" NUMBER(10, 6), "E_GRID_LONG" NUMBER(10, 6), "COMMENTS" VAR"
    "CHAR2(200), "STATE" NUMBER(*,0), "OPER_ID" VARCHAR2(20), "UNQ_ID" VARCHAR2("
    "40), "DESCR" VARCHAR2(200), "S_COUNTY" VARCHAR2(30), "E_COUNTY" VARCHAR2(30"
    ")) PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 STORAGE(INITIAL 65536 FRE"
    "ELISTS 1 FREELIST GROUPS 1) TABLESPACE "USERS" LOGGING"
    Again the next one is
    IMP-00003: ORACLE error 376 encountered
    ORA-00376: file 7 cannot be read at this time
    ORA-01110: data file 7: 'D:\ORACLE\ORADATA\ITMS\USERS01.DBF'
    IMP-00017: following statement failed with ORACLE error 376:
    "CREATE INDEX "TDR_INDX" ON "ITMS_SCHED_TRIPS" ("SEGMENTID" , TRUNC("TRAVELDAT"
    "E") ) PCTFREE 10 INITRANS 2 MAXTRANS 255 STORAGE(INITIAL 65536 FREELISTS 1"
    " FREELIST GROUPS 1) TABLESPACE "INDX" NOLOGGING"
    Thanks in advance.
    Abhijeet

    Hi,
    Please check whether you have permission to write on this 'D:\ORACLE\ORADATA\ITMS\USERS01.DBF' .
    Also please check if that file is actually present there or not...?
    Error ORACLE error 1658 means that either your datafile has filled completely and its not set to auto extend OR hard disk where that file is residing is full and oracle cannot allocate the minimum memory to the segment that is Extent
    Regards

  • Why does iPhoto say "Error writing to disk"?

    When I go on iPhoto I click on the "import all" button and try imorting my pictures from my iPhone to my mac and after a while of importing it completely stops and a window pops out saying "Error writing to disk" and I really don't know what to do to this solution.

    Hello JJ,
    Could be many things, we should start with this...
    "Try Disk Utility
    1. Insert the Mac OS X Install disc, then restart the computer while holding the C key.
    2. When your computer finishes starting up from the disc, choose Disk Utility from the Installer menu at the top of the screen. (In Mac OS X 10.4 or later, you must select your language first.)
    Important: Do not click Continue in the first screen of the Installer. If you do, you must restart from the disc again to access Disk Utility.
    3. Click the First Aid tab.
    4. Select your Mac OS X volume.
    5. Click Repair. Disk Utility checks and repairs the disk."
    http://docs.info.apple.com/article.html?artnum=106214
    Then try a Safe Boot, (holding Shift key down at bootup), run Disk Utility in Applications>Utilities, then highlight your drive, click on Repair Permissions, reboot when it completes.
    (Safe boot may stay on the gray radian for a long time, let it go, it's trying to repair the Hard Drive.)

  • Does /sapmnt need in cluster file system(SAP ECC 6.0 with oracle RAC)

    We are going to be installing SAP with Oracle 10.2.0.4 RAC on Linux SuSE 10 and OCFS2. The Oracle RAC documentation states:
    You must store the following components in the cluster file system when you use RAC
    in the SAP environment:
    - Oracle Clusterware (CRS) Home
    - Oracle RDBMS Home
    - SAP Home (also /sapmnt)
    - Voting Disks
    - OCR
    - Database
    What I want to ask is if I really need put SAP Home(also /sapmnt) on cluster file system? I will build a two nodes oracel 10g RAC and I also have another two nodes to install SAP CI and DI. My orginial think is sapmnt is a NFS share, and mount to all four nodes(RAC node and CI/DI), and all oracle stuff was on OCFS2(only two rac nodes are OCFS), anybody can tell me if SAP Home(also /sapmnt) can be NFS mount not OCFS2, thanks.
    Best regards,
    Peter

    Hi Peter,
    I don't think you need to keep /sapmnt in  ocfs2 . Reason any file system  need  to be in cluster is,in RAC environment, data stored in the cache of one Oracle instance to be accessed by any other instance by transferring it across the private network  and preserves data integrity and cache coherency by transmitting locking and other synchronization information across cluster nodes.
    AS this applies to redo files, datafiles and control files only ,  you should be fine with nfs mount of /sapmnt sharing across and not having ocfs2.
    -SV

  • Using Java with Oracle Spatial

    I am using the JDBC (Java classes dowloaded from
    technet.oracle.com - (sdoapi.zip)) together with Oracle Spatial
    8.1.6. I am writing a collection of Linestrings into Oracle
    Spatial using the supplied java classes(sdoapi.zip). I use the
    function createGeometryCollection(myCollection)to write the
    Linestring collection to Oracle Spatial and they are written to
    the database without error. However, when I try to read the
    object back I get the following error:
    Exception in thread "main"
    oracle.sdoapi.geom.InvalidGeometryException
    at oracle.sdoapi.adapter.AdapterSDO.importGeometry
    (AdapterSDO.java:101)
    at
    MapInfoToOracleSpatial.MapInfoToOracleSpatial.main
    (MapInfoToOracleSpatial.java:211)
    I do not have a listing of this code so can you please tell me
    what the error is (line of code) at line 211. Also, when I use
    the Oracle Spatial Index Advisor to view the spatial data, it
    only displays a quarter of the data. When I use MapInfo to view
    it's equivalent TAB (Table) file it views correctly all the
    data.
    I can't understand why this is happening because the writing to
    oracle and creation of each individual linestring does not
    throw an error. It is only when I try to read back larger
    objects from the database (2500 points) that it gives an error.
    Is there a limitation on the size one can create linestrings?
    Thanks
    Regards
    Stephen Syrett
    Mobile Telecommunications Network
    (MTN) South Africa

    Latest Oracle Openworld Spatial/MapViewer activities list:
    http://www.oracle.com/us/products/database/options/spatial/ssLINK/oow10-focuson-spatial-169204.pdf
    thanks
    LJ

  • Solaris x86 with Oracle RAC 10g Enterprise Edition Release 10.2.0.3.0

    Hello,
    Maybe you can help me (new on RMAN backup) in doing this.
    I have configured a single Oracle 10g database to have backup with RMAN with following steps:
    1. $ mkdir $ORACLE_BASE/rman_scripts
    2. $ mkdir $ORACLE_BASE/logs
    3. $ mkdir $ORACLE_BASE/tracking
    4. $ mkdir $ORACLE_BASE/c_backup
    5. $ sqlplus sys/<password> as sysdba
    6. SQL> alter system set db_recovery_file_dest_size = 50G scope=both;
    7. SQL> alter system set db_recovery_file_dest='${ ORACLE_BASE}/flash_recovery_ area' scope=both;
    8. SQL> alter system set log_archive_dest_10='location= use_db_recovery_file_dest';
    9. SQL> shutdown immediate
    10. SQL> startup nomount
    11. SQL> alter database archivelog;
    12. SQL> alter database open;
    13. SQL> alter database enable block change tracking using file '${ORACLE_BASE}/tracking/rman_ change_track.f';
    14. $ rman target /
    15. RMAN> CONFIGURE CONTROLFILE AUTOBACKUP FORMAT FOR DEVICE TYPE DISK
    TO '/var/opt/oracle/flash_ recovery_area/ORCL/c_backup/% F';
    16. RMAN> CONFIGURE CONTROLFILE AUTOBACKUP ON;
    17. RMAN> CONFIGURE BACKUP OPTIMIZATION ON;
    18. RMAN> CONFIGURE RETENTION POLICY TO RECOVERY WINDOW OF 7 DAYS;
    19. RMAN> exit
    I need to configure incremental backup with RMAN on a two node Solaris x86 with Oracle RAC 10g Enterprise Edition Release 10.2.0.3.0 installation.
    We also use ASM to store database files, and have Oracle software installed on separate file systems (two Oracle roots for Node1 and Node2).
    I have following questions:
    1) where to put Flash Recovery Area (FRA)?
    I saw recommendations to put FRA on the ASM, is this the best way to do it?
    2) Can I put FRA on another file system (not on the ASM) which is available only from Node1? This way I can save space on the ASM.
    3) Is it possible/recommended to run RMAN from Node1 only?
    Below is the script used to run RMAN on the normal Oracle database (without RAC) which I need to change :
    =============================================================================================
    2.0 Oracle backup script: /opt/app/oracle/rman_scripts/backup.sh
    Use this for daily backups, possiblly as a cron job.
    Once a week run this: /opt/app/oracle/rman_scripts/backup.sh FULL
    All other days of the week: /opt/app/oracle/rman_scripts/backup.sh INCREMENTAL
    Note: You may have to change ORACLE_SID, ORACLE_BASE below to match your database.
    =============================================================================================
    #!/usr/bin/ksh
    ORACLE_SID=orcl
    ORACLE_BASE=/opt/app/oracle
    ORACLE_HOME=${ORACLE_BASE}/product/10.2.0/db_1
    PATH=${ORACLE_HOME}/bin:/usr/bin
    LOGDIR=${ORACLE_BASE}/logs
    LOGFILE=${LOGDIR}/rman.log
    if [[ $# < 1 ]]
    then
    echo "usage: backup.sh FULL|INCREMENTAL"
    exit;
    fi
    BACKUPTYPE=${1}
    full='FULL'
    incremental='INCREMENTAL'
    if [[ $BACKUPTYPE == $full ]]
    then
    $ORACLE_HOME/bin/rman target / nocatalog log ${LOGFILE} append << eof
    run {
    backup database;
    SQL 'alter system archive log current';
    backup archivelog all;
    delete noprompt obsolete;
    exit;
    eof
    echo ''
    fi
    if [[ $BACKUPTYPE == $incremental ]]
    then
    $ORACLE_HOME/bin/rman target / nocatalog log ${LOGFILE} append << eof
    run {
    backup database;
    backup incremental level 1 database;
    SQL 'alter system archive log current';
    backup archivelog all;
    delete noprompt obsolete;
    exit;
    eof
    echo ''
    fi

    Hi [email protected],
    Q1) where to put Flash Recovery Area (FRA)?
    A1) With RAC: on the shared storage
    I saw recommendations to put FRA on the ASM, is this the best way to do it?
    If you want your backups to be available for both nodes you have to use shared storage or tape using an mml library.
    So if you want to use the FRA for rman backups and the database is on ASM just make ASM the standard for the FRA as well.
    Q2) Can I put FRA on another file system (not on the ASM) which is available only from Node1? This way I can save space on the ASM.
    A2) Than you cannot recover in case Node1 is down. Best would be to send your storage admin to a training course so he can manage the clustered raw devices needed for ASM.
    Q3) Is it possible/recommended to run RMAN from Node1 only?
    A3) No see A2.
    Regards,
    Tycho

  • Problem with oracle 9i DS with oracle 9i database

    I tried to install Oracle 9i DS on oracle 9i database.
    database successfully installed, but when installing 9i DS, I get error "error in writing to file C:\oracle\Orahome90\ociw32.dll ....
    I retry even after uninstalling all the oracle product, deleted all the oracle keys from the registry and restart the PC.
    I get same result again.
    If I IGNORE and continue with the setup of 9i DS, I am getting following error:-
    while running forms developer -- ifbld90.exe --application failed to start because oci.dll was not found.
    -frm-10142 - HTTP listener was not running on port 8888.
    while running reports developer - rwbuilder.exe -- application failed to start because oci.dll was not found.
    I have tried clean installation after deinstalling again and again, but all in vain.
    I convey my sincere thanks in advance if anybody can suggest solution to my email [email protected]
    bye for now.
    Manoj

    Some hints:
    Check if the listener service is running.
    Play with domain extensions in the *.ora files. This is such a messy part with Oracle. Usually I keep trying several combinations in tnsnames.ora (.world, .<domain>, no domain extension) until it works.
    Sometimes changing "service_name=" to "sid=" in tnsnames.ora works.

  • Problem with ORACLE database

    Hi all,
    I am facing some problem with ORACLE database while configuring one JDBC scenario in the receiver side. I have configured one JDBC scenario like FILE TO JDBC. As JDBC is at receiver side i have provided the input file with the following format ,
    <root>
      <StatementName1>
         <dbTableName action="INSERT">
             <table>employee</table>
              <access>
                    <flag>NO</flag> 
                    <ID>567</ID>
             <NAME>134</NAME>
                 </access>
         </dbTableName>
      </StatementName1>
    </root>
    But in the ORACLE database all my columns are in the uppercase.
    So when i have tried to send this input file to process through JDBC receiver side then i have faced the below error :
    FATAL ERROR: Column 'flag' does not exist in table 'employee'
    Then I have tried to check with sender side processing of JDBC adapter. By using the SQL statements as
    SELECT * FROM EMPLOYEE WHERE flag='NO'
    UPDATE EMPLOYEE SET Flag='YES'  WHERE flag='NO'
    Here it worked fine.
    Now my question is,
    If the same adapter is working when we are writing a query directly(sender side), then we must also allow case independence in the auto generated SQL part(receiver side).
    The JDBC drivers for both SQL server and Oracle, supports the same java code, for accessing the DB, irrespective of the case of the column names.
    Please clarify this doubt as soon as possible.
    Thanks,
    Soorya.

    Hi Dharmaveer,
    When I go with the following input file,
    <root>
      <StatementName1>
         <dbTableName action="INSERT">
             <table>EMPLOYEE</table>
              <access>
                    <FLAG>NO</FLAG> 
                    <ID>666</ID>
                    <NAME>SSS</NAME>
                 </access>
             </dbTableName>
      </StatementName1>
    </root>
    I will get the following query with succesfull processing.
    INSERT INTO  EMPLOYEE (FLAG, ID, NAME) VALUES (NO, 666, SSS)
    when i go with this input file,
    <root>
      <StatementName1>
         <dbTableName action="INSERT">
             <table>EMPLOYEE</table>
              <access>
                                    <flag>NO</flag> 
                                   <ID>777</ID>
                        <NAME>TTTT</NAME>
                        </access>
                   </dbTableName>
      </StatementName1>
    </root>
    Then i am facing this problem,
    Message processing failed. Cause: com.sap.aii.af.ra.ms.api.RecoverableException: Error processing request in sax parser: Error when executing statement for table/stored proc. 'EMPLOYEE' (structure 'StatementName1'): java.sql.SQLException: FATAL ERROR: Column 'flag' does not exist in table 'EMPLOYEE'
    Here I cant get any SQL statement as it is not successfull.
    I have gone through your link but it also says to modify the column in the input file which is not possible for me. Can you please help me in this regard?
    Thanks,
    Soorya

Maybe you are looking for

  • Work Order - TECO Status Question

    Dear All, Is there a way to indicate who performed the last accounting transaction with a Work order? For example, Who Teco'd the order and when it was Teco'd, etc.. This allows me to create a report to show productivity of how many work orders being

  • Pls help me get itunes back on my computer.  runtime error R6034 and error 7 windows error 1114

    when i updated my itunes i got an error that it couldnt install it and to do it manually which didnt work either so i tried the uninstall and reinstall but i keep getting a runtime error R6034 and error 7 windows error 1114 and it wasnt installed cor

  • Queues in READY status

    Dear All, This question has been posted many times but never saw a conclusive answer on this forum. To repeat, all settings of scheduler being fine (SMQS - Type R), there are large number of queues awaiting to be manually unlocked from READY status t

  • Changing currency and price on BAPI_PO_CHANGE

    I gurus! I need change the currency with BAPI_PO_CHANGE, the first time i´ts works but when i want change another time, the net_price and gr_price is wrong. I change from MXN to USD, later i change from USD to MXN. any Idea?

  • Script logic: differences between MS and NW?

    Hi to all, at last SAP Tech Ed I've bought a book on SAP BPC 7, Netweaver version...I did it because I'm already working with the MS version, but I hoped to get a better vision on BPC over NW, and to understand some critical subjects. One question th