Times Ten and PL/SQL XML Type

How TT can support XML type PL/SQL or what is workaround...

Gennady Sigalaev wrote:
PL/SQL in TimesTen does not support large objects (LOBs), Internet data types (XMLType, URIType, HttpURIType), or "Any" data types (AnyType, AnyData, AnyDataSet).
More information you can find in documentation (http://download.oracle.com/docs/cd/E13085_01/welcome.html)
From Oracle's website: http://www.oracle.com/technetwork/database/timesten/documentation/1121-historic-183693.html
"You can cache Oracle LOB data in TimesTen cache groups. Oracle CLOB data is cached as TimesTen VARCHAR2 data. Oracle BLOB data is cached as TimesTen VARBINARY data. Oracle NCLOB data is cached as TimesTen NVARCHAR2 data."
However, the catch is the VARCHAR2 and VARBINARY in TT would appear to only handle 4 MB and Oracle CLOB and BLOB handles 4 GB. So it does support it but up to a certain point it would seem.
Edited by: victor on Aug 24, 2011 12:36 PM

Similar Messages

  • Info about Time, GIS and other usable XML(DB) Services

    Based on Re: how to retrive World Server time in Sql here some thoughts about "TimeService the XMLDB Way".
    This post for all who are in need of such a solution. If I have time I will also post the proper XMLTABLE alternative.
    Probably there are better ways - Does anyone know better / useful web services? (that can be used with or called from XMLDB).
    Cool XMLDB examples that can be used by others are appreciated.
    ...while thinking about it I would do it for instance via HTTPURITYPE (also based on UTL_HTP btw).
    In "steps" (based on XMLDB functionality and http://www.nanonull.com/TimeService/TimeService.asmx)
    SQL> set long 1000000
    SQL> set pages 50000
    SQL> select httpuritype('http://www.nanonull.com/TimeService/TimeService.asmx/getTimeZoneTime?timezone=CET').getClob()
      2  from dual;
    HTTPURITYPE('HTTP://WWW.NANONULL.COM/TIMESERVICE/TIMESERVICE.ASMX/GETTIMEZONETIM
    <?xml version="1.0" encoding="utf-8"?>
    <string xmlns="http://www.Nanonull.com/TimeService/">5:54 PM</string>
    SQL> select extract(xmltype(httpuritype('http://www.nanonull.com/TimeService/TimeService.asmx/getTimeZoneTime?timezone=CET').getCLOB())
      2                      ,'/*','xmlns="http://www.Nanonull.com/TimeService/"')
      3  as "CET Time"
      4  from dual;
    CET Time
    <string xmlns="http://www.Nanonull.com/TimeService/">6:00 PM</string>
    SQL> select extractvalue(xmltype(httpuritype('http://www.nanonull.com/TimeService/TimeService.asmx/getTimeZoneTime?timezone=CET').getCLOB())
      2                      ,'/string','xmlns="http://www.Nanonull.com/TimeService/"')
      3        as "CET TIME"
      4  from dual;
    CET TIME
    6:01 PM
    SQL> create table M as
      2  select extractvalue(xmltype(httpuritype('http://www.nanonull.com/TimeService/TimeService.asmx/getTimeZoneTime?timezone=CET').getCLOB())
      3                      ,'/string','xmlns="http://www.Nanonull.com/TimeService/"')
      4        as "CETTIME"
      5  from dual;
    Table created.
    SQL> desc M
    Name                                      Null?    Type
    CETTIME                                            VARCHAR2(4000)The "Time" in London (one of the strings that worked... LOV for the city's would be useful...?)
    SQL> select extractvalue(xmltype(httpuritype('http://www.nanonull.com/TimeService/TimeService.asmx/getCityTime?city=London').getCLOB())
      2                      ,'/string','xmlns="http://www.Nanonull.com/TimeService/"')
      3        as "The Time in London"
      4  from dual;
    The Time in London
    5:09 PM

    XML Schema: http://www.hostip.info/api/hostip-1.0.0.xsd
    SQL> SELECT *
      2  FROM v$version;
    BANNER
    Oracle DATABASE 11g Enterprise Edition Release 11.1.0.6.0 - Production
    PL/SQL Release 11.1.0.6.0 - Production
    CORE    11.1.0.6.0      Production
    TNS FOR 32-bit Windows: Version 11.1.0.6.0 - Production
    NLSRTL Version 11.1.0.6.0 - Production
    5 rows selected.
    SQL>
    SQL> col Name          FOR a20
    SQL> col CountryName   FOR a20
    SQL> col CountryAbbrev FOR a20
    SQL> col Coordinates   FOR a20
    SQL>
    SQL>
    SQL> SELECT extractvalue(value(t),'/Hostip/gml:name'
      2                          , 'xmlns:gml="http://www.opengis.net/gml"
      3                             xmlns="http://www.hostip.info/api"')   AS "Name"
      4  ,      extractvalue(value(t),'/Hostip/countryName'
      5                          , 'xmlns:gml="http://www.opengis.net/gml"
      6                             xmlns="http://www.hostip.info/api"')   AS "CountryName"
      7  ,      extractvalue(value(t),'/Hostip/countryAbbrev'
      8                          , 'xmlns:gml="http://www.opengis.net/gml"
      9                             xmlns="http://www.hostip.info/api"')   AS "CountryAbbrev"
    10  ,      extractvalue(value(t),'/Hostip/ipLocation/gml:PointProperty/gml:Point/gml:coordinates'
    11                          , 'xmlns:gml="http://www.opengis.net/gml"
    12                             xmlns="http://www.hostip.info/api"')   AS "Coordinates"
    13  FROM TABLE (XMLSEQUENCE (EXTRACT (HTTPURITYPE ('http://api.hostip.info/?ip=141.146.8.66').getxml ()
    14                           , '/HostipLookupResultSet/gml:featureMember/Hostip'
    15                           , 'xmlns:gml="http://www.opengis.net/gml"
    16                              xmlns="http://www.hostip.info/api"' ))) t
    17  /
    Name                 CountryName          CountryAbbrev        Coordinates
    Redwood City, CA     UNITED STATES        US                   -122.206,37.5164
    1 row selected.

  • How to combine "Object-to-XML (OXM)" and "Direct to XML Type" mapping?

    hi
    If I have an XMLType column in my table (wich I can map using TopLink) and I have defined the structure of the contents of this XMLType column using XML Schema (wich I can map using Toplink), how can I combine both types of TopLink mappings "as transparently as possible"?
    for "Object-to-XML (OXM)" mapping
    see http://www.oracle.com/technology/products/ias/toplink/technical/tips/ox/index.htm
    for "Direct to XML Type" mapping
    see http://www.oracle.com/technology/products/ias/toplink/doc/1013/main/_html/relmapun004.htm#CHDFIFEF
    thanks
    Jan Vervecken

    Thanks for your reply James Sutherland.
    Although I haven't used a "TopLink Converter" before, this seems like a good idea.
    The thing is that the "TopLink Workbench Editor" for my "Direct to XML Type" mapping doesn't have a "Converter" tab, some other mapping type editors do have such a "Converter" tab.
    I'm not sure if I completely understand how such a "TopLink Converter" is supposed to work. How many attributes do I need in the "XMLRow" Java object for the "MY_XML" column in the "XML_TABLE" table I try to map to?
    I suppose I should try to get a situation where the "XMLRow" Java object has an "myXML" attribute of Java class type "MyXML" (where "MyXML" has been mapped to an XML Schema), not?
    So do I also still need an attribute "myXMLDocument" of type org.w3c.dom.Document as I do now for the "Direct to XML Type" mapping?
    Oh, by the way ... for anyone who hits this forum thread looking for the reason why the TopLink Workbench reports the problem "Attribute must be assignable to java.lang.String, org.w3c.dom.Document, or org.w3c.Node" while your attribute is of such a type, read this forum post
    Re: Toplink WB 10.1.3 - Aggregate field mapping bug and XMLType question
    For me the "Direct to XML Type" mapping works fine, just ignoring the waring. This is supposed to be bug number 5071250.
    thanks
    Jan Vervecken

  • Times-Ten and OracleText Queries

    Can Times-Ten be used in conjunction with CONTAINS queries? We need to cache the results of some OracleText queries.

    The TimesTen database doesn't support CONTAINs queries and there is no integration with OracleText at this point.
    - scheung

  • Binding ADF UI to a complex XML Type – Development Productivity, State Mgmt

    Binding ADF UI to a complex Web Service – Development Productivity, Managing State, Etc.
    For those scenarios when ADF UI has to be bound to a complex XML type, it seems that the direction taken by the ADF development team was to generate data-controls that map to complex XML types.
    The major problem with this approach is the fact that application server does not manage state. For example, If application has to submit a purchase order to a web service then application server has to manage the shopping card state until its ready to be submitted.
    In some previous forum threads it had been suggested that using generated data-controls is a “quick and dirty” solution for accessing web services and that a proper solution is to use generated Java Proxy Interface.
    That implies many manually constructed entity objects, associations and overridden doDML methods, which is far less productive compared to entity objects generated from a database schema.
    Suggestion and a Question for the ADF development team
    From the logical model standpoint, there is no difference in between the XML schema and DB schema. Therefore, it is possible to generate entity objects and associations for XML Types of XML schema the same way it’s done for tables of Database schema. The only difference is serialization of CRUD operations on entity objects.
    One way to handle this is:
    -     Generate Application Module for a complex XML type. This application module should have methods for marshaling to/from corresponding XML type.
    -     Generate entity object for each subtype within the complex XML type. In case of nested subtypes, generate association to a parent type.
    -     Generate data control with actions that are bound to web service operations and code that serializes request message from corresponding AM and de-serializes response message to corresponding AM.
    This way, ADF would offer the same productivity for the SOA development as one its currently offering for the ORM development.
    Until the time when something like this is available, what would be the best approach for binding ADF UI to web services?
    Feedback is greatly appreciated.
    Boro Petrovic
    Edited by: wds12518 on Jan 25, 2010 11:49 AM

    We have similar issues as our big portion of the UI is based on WS. We found that there is no easy way to map entity object structure to complex XML type (one EO can based on one flat type or domain can't be bound to UI directly). Oracle PMs, is there any better solution or future plans to address this issue?

  • Client Server mode for time ten connection

    Hi, I am new to Times Ten and got a question for running TimesTen in Client Server Mode. When it is running in Client/Server mode:
    Does the database just run like a regular remote database ?
    Is there security for the remote application, such as userid/password ?
    Is there any limitation on number of remote connection can concurrently connect to the DB ?
    Does it use the JDBC or it has to be the native library ?
    Thank you for your help
    Henry

    Hi,
    With regard to your questions regarding TimesTen client/server mode:
    1. Does the database just run like a regular remote database ?
    I'm not sure exactly what you are getting at here but yes, I would say that it runs like any other c/s database. If you can be more specific in your question then I can give a more precise answer.
    2. Is there security for the remote application, such as userid/password ?
    If the TimesTen instance to which you are connecting was installed with Access Control enabled then yes, normal userid/password security applies. The client application must provide a valid TimesTen userid and password in order to establish a connection.
    3. Is there any limitation on number of remote connection can concurrently connect to the DB ?
    Subject to O/S resource limitations, a single instance of TimesTen (i.e. the TT server) supports a maximum of 9,999 concurrent client/server connections. Each individual datastore supports a maximum of 2047 concurrent connections (direct and client/server).
    4. Does it use the JDBC or it has to be the native library ?
    The TimesTen JDBC driver supports both direct modde and client/server. The TimesTen ODBC driver is provided in both direct mode and client/server versions. The JDBC driver uses the ODBC driver (it is a type 2 JDBC driver) and so the client machine requires a TimesTen client installation to use either ODBC or JDBC to access a remote datastore.
    Chris

  • Caching configuration (Times Ten - Oracle 10g)

    Hi all,
    I have managed to setup Times Ten and get a cache group configured against a table in an oracle 10g instance.
    That all works so great. My question is how do I define a set of cache groups and get Times Ten to load them in automatically all together. So what I am saying, is I dont want to have to enter these cache group definitions in manually via ttisql command line (as I have done this time).
    Could someone please point me in the right direction with this? Also I would like to be able to set TT up such that the cache agent is automatically started. Is this possible?
    Thanks,
    Dan

    Hi Dan,
    In the same way that you can create cache groups through a ttIsql script you can also load them from a script using the LOAD CACHE GROUP and REFRESH CACHE GROUP commands.
    However, let's remember that TimesTen is a persistent database and so you should not need to re-create the cache groups nor re-load them just becasue you have stopped and started TimesTen. The cache group definitions and data will still be present as they are persistently stored. Also, if you have a READONLY or USERMANAGED cache group using AUTOREFRESH then the changes that occur in oracle are captured and refreshed to the cache group(s) automatically.
    As for starting the cache agent 9or the repagent0 automatically, this kind of depends on how you are managing the startup/shutdown of the cache datastore (i.e. what itss ramPolicy is). You can set the cachePolicy for the datastore using ttAdmin as follows:
    ttAdmin -cachePolicy always DSNname
    this tells TimesTen that the cache agent should always be running for this datastore. However, a side effect of this is that the datastore will allways be loaded in memory when the TimesTen instance is up (i.e. the TimesTen main daemon is running). If you want to be able to start and stop the datastore manually (or from a script) then you will likely be using a ramPolicy of manual and running some script to load/unload the datastore into/from memory. If so then just add the relevant ttadmin -cacheStart|-cacheStop commands to that script.
    Chris

  • XML Publisher(XDODTEXE) in EBS taking more time with the same SQL in TOAD

    HI
    XML Publisher(XDODTEXE) in EBS taking more time with the same SQL in TOAD.
    The sql has 5 union clauses.
    It takes 20-30 minutes in TOAD compared to running through Concurrent Program in XML Publisher in EBS taking around 4-5 hours.
    The Scalable Flag at report level is turned on with the JVM options set to -Xmx1024m -Xmx1024m in Concurrent Program definition.
    Other configurations for Data Template like XSLT, Scalable, Optimization are turned on though didn't bounce the OPP Server for these to take effect as I am not sure whether it is needed.
    Thanks in advance for your help.

    But the question is that how come it is working in TOAD and takes only 15-20 minutes?
    with initialization of session ?
    what about sqlplus ?
    Do I have to set up the the temp directory for the XML Publisher report to make it faster?
    look at
    R12: Troubleshooting Known XML Publisher and E-Business Suite (EBS) Integration Issues (Doc ID 1410160.1)
    BI Publisher - Troubleshooting Oracle Business Intelligence (XML) Publisher For The Oracle E-Business Suite (Doc ID 364547.1)

  • Unable Update SQL Developer with Times Ten in-Memory DB Extension 1.2.1.1.0

    SQL Developer 1.2.1 Build MAIN 32.13 (Windows XP)
    We can't Update SQL Developer with Times Ten in-Memory DB Extension 1.2.1.1.0 :-((
    We did the same actions as described in tutorial
    http://www.oracle.com/technology/products/timesten/viewlets/tt703_sqldev_install_ext_viewlet_swf.html
    In addition to tutorial we have "Log In" window
    "To download Oracle Times Ten in-Memory Database Extension
    you must enter your Oracle Web Account user name and password"
    This window was not displayed in tutorial...
    We try to use three our OTN accounts and the result
    was the same - we was asked with "User name" and "Password"
    again, again, again and again,,, :-(
    Is it a bug or a feature?
    Mayby we didn't understand something or did something wrong?

    Hi Simon,
    I found cause of update failure! It was the HTTP proxy setting.
    I turned off "Use HTTP proxy server" checkbox in SQL Developer -> Tools -> Preferences -> Web Browser and Proxy.
    After this action SQL Developer was successfully updated with Times Ten Extension.
    Valery Yourinsky

  • PLSQL and Java XML type mapping

    Running database 10gR2.  Trying to create a Java stored procedure that returns an XML type (oracle.xdb.XMLType) like so:
    class MyCoolClass
       static oracle.xdb.XMLType doSomething( String xmldata)
         oracle.xdb.XMLType xmlt = new oracle.xdb.XMLType( xmldata );
         return xmlt;
    This, of course, is wrapped by a PL/SQL function:
    function doSomething( xmldata varchar2 ) return SYS.XMLTYPE AS LANGUAGE JAVA
    name 'MyCoolClass.doSomething(java.lang.String) return oracle.xdb.XMLType';
    While everything compiles (that means it should work, right?), the above codes generates a runtime error ORA-00932 'inconsistent datatypes' error.   I'm not  sure what I'm doing wrong, it was my understanding that a default mapping existed between the PL/SQL SYS.XMLTYPE and oracle.xdb.XMLType.  Is what I'm attempting even possible?

    Sorry, I was being sloppy with my code example.  It should be:
    class MyCoolClass
       static oracle.xdb.XMLType doSomething( String xmldata)
          java.sql.Connection c = java.sql.DriverManager.getConnection("jdbc:default:connection:");
         oracle.xdb.XMLType xmlt = new oracle.xdb.XMLType( c, xmldata );
         return xmlt;

  • SQL/XML performance and xdbcore-parameter

    Hello,
    I try to figure out, what the best method to retrieve large XML-Documents is.
    I compare performance and resources needed for: SQL/XML, DBMS_XMLGEN, SYS_XMLGEN and XSU.
    With a fixed set of 1000 rows I created XML of about 32 MB.
    I am on 10.2.0.4 on Linux
    XML is not schema based.
    SQL*Plus: Release 10.1.0.4.2 - Production on Fri Aug 21 13:35:42 2009
    Copyright (c) 1982, 2005, Oracle.  All rights reserved.
    Connected to:
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - Production
    With the Partitioning, Data Mining and Real Application Testing optionsWhat I see, is that SQL/XML is the slowest and needs huge PGA amount (160 Mb and seems to increase linearly, for 5000 rows I've measured ~800 Mb).
    As I've read here, memory issues with SQL/XML should have been solved in 9.2.0.3 (?)
    Is this the case? Or is this PGA amount ok?
    I would like also to test the impact of xdbcore-xobmem-bound and xdbcore-loadableunit-size parameters.
    But I cannot observe any impact. Must I restart the database everytime after change of xdbconfig.xml?
    But also after increasing from 1024 16 to resp. 10240 5120 and restart the same speed and memory consumption are in place.
    Am I doing something wrong?

    SQL*Loader: Release 11.2.0.2.0 - Production on H. Febr. 11 18:57:09 2013
    Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved.
    Control File: prodxml.ctl
    Character Set UTF8 specified for all input.
    Data File: test.xml
    Bad File: test.bad
    Discard File: none specified
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 5000
    Bind array: 64 rows, maximum of 256000 bytes
    Continuation: none specified
    Path used: Conventional
    Table PRODUCT_XML, loaded from every logical record.
    Insert option in effect for this table: APPEND
    Column Name Position Len Term Encl Datatype
    COL_ID FIRST 1000 CHARACTER
    (FILLER FIELD)
    IN_FILE DERIVED * EOF CHARACTER
    Static LOBFILE. Filename is bv_test.xml
    Character Set UTF8 specified for all input.
    SQL*Loader-605: Non-data dependent ORACLE error occurred -- load discontinued.
    ORA-01652: unable to extend temp segment by 128 in tablespace TEMP
    Table PRODUCT_XML:
    0 Rows successfully loaded.
    0 Rows not loaded due to data errors.
    0 Rows not loaded because all WHEN clauses were failed.
    0 Rows not loaded because all fields were null.
    Space allocated for bind array: 256 bytes(64 rows)
    Read buffer bytes: 1048576
    Total logical records skipped: 0
    Total logical records rejected: 0
    Total logical records discarded: 0
    Run began on H. Febr. 11 18:57:09 2013
    Run ended on H. Febr. 11 19:20:54 2013
    Elapsed time was: 00:23:45.76
    CPU time was: 00:05:05.50
    this is the log
    i have truncated everything i am not able to load 400 mega into 4 giga i cannot understand
    windows is not licensed 32 bit

  • Custom action with XML type input and output parameter.

    Hi,
    I want to develop custom action with xml type input and/or output parameter.
    Is there sample code for java side. How is the definition of input and/or output parameter and set/get methods?
    does it need special .jar file to develop custom action like this?
    Thanks.

    Cemil - yes, you can use XML data types.  Use the class
    com.sap.lhcommon.xml.XMLDataType
    for your parameter type.  Here is a snippet from a custom action we use to log XML (instead of just returning the #text node like the default logger does):
    public class XMLLogger extends ActionReflectionBase
        private String source;
        private String eventType;
        private String textMessage;
        private XMLDataType xmlMessage;
        public XMLLogger()
            log = new Logger("UserLog");
            source = DEFAULT_SOURCE;
            eventType = TYPE_INFO;
            textMessage = "";
            xmlMessage = new XMLDataType();
        public XMLDataType getXmlMessage()
            return xmlMessage;
        public void setXmlMessage(XMLDataType xmlMessage)
            this.xmlMessage = xmlMessage;
        public void Invoke(Transaction transaction, ILog ilog)
            StringBuffer sb = new StringBuffer();
            sb.append('[');
            sb.append(source);
            sb.append("] ");
            sb.append(textMessage);
            sb.append(XMLUtils.convertXmlToString(xmlMessage));
    XMLUtils is a helper class we wrote - it's just a bunch of standard Java XML boilerplate code.  The important part you need to know is XMLDataType.getDocument() will return an org.w3c.dom.Document.
    I hope that was enough information to help.
    -tim

  • Is there any downside to using a reference type library if I use Time Capsule and will do file management only from within Aperture?

    Have done a lot of reading to get prepared to convert to Aperture 3 and have this question regarding setting up my library type.
    I'm not a professional -just a heavy user hobbyist.
    It appears that the major factors in using reference style is to backup the images with Time Capsule and always do moves or deletes from within the Aperture application.
    If my architecture matters here's what I have that may be involved in photo management and editing:
    iMac (latest 27" high power version with lots of memory)
    Internal 2TB HD where the library is stored
    External (FireWire) 2TB HD where the images are stored
    External (FireWire) 6TB HD backup drive
    In the future an iPad for remote work on images (when/if available) and a Mac laptop (for same remote use)
    I see major downside issues to letting the library manage my files - such as inaccessibility (or awkward accessibility) to the images for other programs, and performance issues when the library gets large (thousands of images or 100+ gb in size)
    The chief complaints I've seen regarding using a reference mode is the broken link issues created if file management is done outside of Aperture (adds, deletes, moves of files) and the inability to use the Vault function for backup.
    One feature that I imagine I'd like to have is maximum integration with Final Cut Pro X and that's one area I haven't seen much on and would be interested to hear about if that integration is affected with the choice of managed vs referenced library types. (I like to produce film clips that are combinations of pics and video and want to be set up so that is done in the easiest fashion when working in FCP X)
    I'm sure I've not seen all sides of this issue and would like to see some discussion around this question.
    Thanks to everyone contributing!
    Craig

    Goody, Goody, you hit a few of my favorite subjects! Herewith some comments, with the usual caveats - true to the best of my knowledge and experience, others may have different results, YMMV, and I could be wrong.
    I run Aperture on two machines:
    -- 2007 Mac Pro with 2x2.66 GHz Xeon, 21 GB of RAM, a 5770 GPU, and multiple HD. I have about 11,000 images, taking up 150 GB. (Many are 100 MB TIFF, scanned slides)
    -- 2006 Mac Book Pro with 1 2.0 GHz Core Duo, 2 GB of RAM and a 240 GB SSD.
    Two very different machines.
    -- Aperture Libraries are all the same - Managed or Referenced. If Managed, then the Master image files are inside the Package, if Referenced, they are outside, and you can have any combination you want. Managed is easier, but Referenced is not hard if you are the least bit careful.
    -- While the sheer size of an Aperture Library is not a big issue, the location on disk of the different components can have a tremendous impact on performance.
    -- Solid State Drives (SSD) read and write faster than regular Hard Disks (HD) and, what is more important, empty HD read and write much faster than full HD.
    -- Aperture speed requires a combination of RAM, CPU, graphics processing unit (GPU) speed, and disk speed. The more RAM you have, the less paging you will see. With enough RAM, the next bottle neck is CPU (speed and cores) and GPU speed. But even then you will still have to fetch an image (longer if you pull the full resolution Master, read and rewrite the Version file, and update the Preview and Thumbs.
    So what works?
    -- RAM, RAM, and more RAM. 4 GB will work, but you will page a bit. 8 GB is much faster. On my old MacPro the sweet spot was about 16 GB of RAM.
    -- Keep your Library on your fastest (usually internal) drive. Keep that disk as unloaded as possible. How do you keep it unloaded? Either buy BIG, 1 TB+ or move your Masters off onto another disk. The good news here is that as Master files are written only once and never rewritten, speed of this disk, as opposed to the disk that holds the Library, is not important. There is a one second pause as the Master is read into memory and, if you have enough RAM, that is it - the Master will never be paged out. If, on the other hand, you do not have enough RAM, and you do a lot of adjusting at full resolution, then the speed of the disk that holds your Masters will become very important due to paging.
    -- I found that I picked up a tiny bit of speed by keeping the Masters on a dedicated disk. Thus, in your configuration, if you can dedicate that 2 TB FW HD to your Masters, you should see very nice performance.
    Final notes on backups and archives:
    -- One conventional wisdom is that you should make an archive copy of every image file before or as you load them into your system. (Aperture in this case.) This archive is then never touched or deleted.
    -- An alternative approach is that you do not keep such an archive, but only the images that you have in your Aperture Library. And when you delete from the Library, you no longer keep a copy anywhere.
    I do the following:
    -- Card to Aperture. Card is then kept at least 24 hours until all of my backups have run. (I use three - Time Machine, Clone, and off site.)
    -- I do not keep archive copies. If I decide to delete a file, my only recourse is to recover it from Time Machine during the six month cycle of my Time Machine backups. Thereafter, it is lost.
    There are merits to both approaches.
    Hope this is clear, correct, and responsive to your needs.
    DiploStrat

  • I have been working on the same numbers file for the past few weeks.  The last time I opened it was 1 week ago.  Today when I tried to open it I am unable and getting a message that the file is invalid and the index.xml file is missing.

    I have been working on the same numbers file for the past few weeks.  The last time I opened it was 1 week ago.  Today when I tried to open it I am unable and getting a message that the file is invalid and the index.xml file is missing. 

    Hi Tracie,
    I upgraded to Maverick OS X 10.9.5, numbers spreadsheet is saved. Upon re-opening, it appears to be frozen, a warning "file is invalid as index.xml file is missing". I checked, and the file is not "locked". This appears to occur only with using the new numbers app. When I open previous spreadsheets from old iWorks, no such problem occurs.
    How did you resolve your problem?
    Would appreciate any help here.
    Thanks,
    Deehay

  • Parsing the data from and xml type field

    Hi - I have registered a schema and inserted arecord into the table with the xml type column. Now I want to parse the data from the xmltype field into a relational table. I have been using the following select statement to accomplish this - and it does work if there is data in all the selected fields but when the filed is null then the whole select statement fails and brings back 'no rows returned'.If the value is null I want the select statment to return null. please give any ideas.
    SELECT version,frmd_transaction_date,extractValue(value(b), 'event_update/location')"location",
    extractValue(value(b), 'event_update/sending_system')"sending_system",
    extractValue(value(b), 'event_update/event_identifier')"event_identifier",
    extractValue(value(b), 'event_update/event_link')"event_link",
    extractValue(value(b), 'event_update/organization_code')"organization_code",
    nvl(extractValue(value(c), '/schedule/event_duration_minutes'),'000')"event_minutes"
    FROM fraamed_user.frmd_event_update , TABLE(xmlsequence(extract(xml_event_update, '/event_update')))b,
    TABLE(xmlsequence(extract(xml_event_update, '/event_update/schedule')))c

    ...then I guess you have to rewrite the query.
    Is schedule another xml sequence inside of event_update sequence ?
    If it is not you can try this :
    SELECT version,frmd_transaction_date,extractValue(value(b), '/event_update/location/text()')"location",
    extractValue(value(b), '/event_update/sending_system/text()')"sending_system",
    extractValue(value(b), '/event_update/event_identifier/text()')"event_identifier",
    extractValue(value(b), '/event_update/event_link/text()')"event_link",
    extractValue(value(b), '/event_update/organization_code/text()')"organization_code",
    extractValue(value(b), '/event_update/schedule/event_duration_minutes/text()')"event_minutes"
    FROM fraamed_user.frmd_event_update , TABLE(xmlsequence(extract(xml_event_update, '/event_update')))b
    ...if yes, did you try nvl function (I don't think this would be a solution of a problem):
    SELECT version,frmd_transaction_date,nvl(extractValue(value(b), '/event_update/location/text()')"location", 'NULL VALUE'),
    nvl(extractValue(value(b), '/event_update/sending_system/text()')"sending_system",'NULL VALUE'),
    nvl(extractValue(value(b), '/event_update/event_identifier/text()')"event_identifier",'NULL VALUE'),
    nvl(extractValue(value(b), '/event_update/event_link/text()')"event_link",'NULL VALUE'),
    nvl(extractValue(value(b), '/event_update/organization_code/text()')"organization_code",'NULL VALUE'),
    nvl(extractValue(value(c), '/schedule/event_duration_minutes/text()')"event_minutes",'NULL VALUE')
    FROM fraamed_user.frmd_event_update , TABLE(xmlsequence(extract(xml_event_update, '/event_update')))b,
    TABLE(xmlsequence(extract(xml_event_update, '/event_update/schedule')))c
    If none of this works post your xml schema.

Maybe you are looking for