Parsing and Indexing

I am writing a PL/SQL routine to parse, identify individual words from a string (NVARCHAR2/NCLOB) and store them as individual words in an "index" table so they can be searched upon later. My lack of experience in NLS shows here. My question is...When parsing the NVARCHAR2/NCLOB strings how do I identify which characters should be part of a word and which are expendable (punctuation, etc). Do I have to store and manage characters from each locale and then identify the local prior to parsing and indexing? This seems like an incredible amount of work not to mention needing somebody that understands each locale. Any help is appreciated.
Thanks,
Greg
null

This seems like an incredible amount of work not to mention needing somebody that understands each localeYes it is. Thats why Oracle has a product called Oracle Text. Oracle Text uses standard SQL to index, search, and analyze text and documents stored in Oracle8i, in files and on the Web. It also indexes metadata extracted from image, audio, and video media by Oracle interMedia. It supports 39 langauges.
Check out http://technet.oracle.com/products/intermedia/ and click Oracle Text for more information
null

Similar Messages

  • Error parsing document index: invalid character in attribute value

    Hi,
    I am working on an eBook with iBook Author. Once in a while, after I finish working, and want to reopen my eBook a few hours or days after I get the error:
    'title' could not be opened
    Error parsing document index: invalid character in attribute value
    I have to go back to an old backup I always do before quitting.
    What is the reason of this error message and anyw ay to fix/repair it on a version I just saved?
    Thanks,
    JP.

    Hello,
    I have the same error, but for me the above solution did not work.
    Did I understand correctly? This is what I did:
    1. I changed the extension of the IBA file in ZIP
    2. I unzipped the file
    3. in the folder with the unzipped book I have changed the file index.xml in index.html
    4. I zipped it all back (in a ZIP file)
    5. I renamed the extension ZIP archive in IBA
    6. I tried to open the book and I got  the error that there was no index.xml file
    7. I changed the extension of the IBA and in ZIP
    8. I unzipped the file again
    9. I changed back index.xml to index.html and I zipped it back (compressed it)
    10. I renamed the extension ZIP archive in IBA
    11. I opened the book
    Is this correct?
    The problem is that I am still receiving the same error message: Error parsing document index: invalid character in attribute value
    Did I do something wrong?
    If you can help I would be very greatfull, I worked hard on this book and it is the only backup that I have saved.
    Thank you!

  • Goldengate Extracts reads slow during Table Data Archiving and Index Rebuilding Operations.

    We have configured OGG on a  near-DR server. The extracts are configured to work in ALO Mode.
    During the day, extracts work as expected and are in sync. But during any dialy maintenance task, the extracts starts lagging, and read the same archives very slow.
    This usually happens during Table Data Archiving (DELETE from prod tables, INSERT into history tables) and during Index Rebuilding on those tables.
    Points to be noted:
    1) The Tables on which Archiving is done and whose Indexes are rebuilt are not captured by GoldenGate Extract.
    2) The extracts are configured to capture DML opeartions. Only INSERT and UPDATE operations are captured, DELETES are ignored by the extracts. Also DDL extraction is not configured.
    3) There is no connection to PROD or DR Database
    4) System functions normally all the time, but just during table data archiving and index rebuild it starts lagging.
    Q 1. As mentioned above, even though the tables are not a part of capture, the extracts lags ? What are the possible reasons for the lag ?
    Q 2. I understand that Index Rebuild is a DDL operation, then too it induces a lag into the system. how ?
    Q 3. We have been trying to find a way to overcome the lag, which ideally shouldn't have arised. Is there any extract parameter or some work around for this situation ?

    Hi Nick.W,
    The amount of redo logs generated is huge. Approximately 200-250 GB in 45-60 minutes.
    I agree that the extract has to parse the extra object-id's. During the day, there is a redo switch every 2-3 minutes. The source is a 3-Node RAC. So approximately, 80-90 archives generated in an hour.
    The reason to mention this was, that while reading these archives also, the extract would be parsing extra Object ID's, as we are capturing data only for 3 tables. The effect of parsing extract object id's should have been seen during the day also. The reason being archive size is same, amount of data is same, the number of records to be scanned is same.
    The extract slows down and read at half the speed. If normally it would take 45-50 secs to read an archive log of normal day functioning, then it would take approx 90-100 secs to read the archives of the mentioned activities.
    Regarding the 3rd point,
    a. The extract is a classic extract, the archived logs are on local file system. No ASM, NO SAN/NAS.
    b. We have added  "TRANLOGOPTIONS BUFSIZE" parameter in our extract. We'll update as soon as we see any kind of improvements.

  • Parse and render XML... and get original replicated?

    Now that I've (temporarily) got iFS working (on another machine for now). I've got an XML question: I moved the airportdefinition.xml into the folder, and then subsequently moved the airport.xml into the folder, whereupon iFS parsed it and stored it. Yeah! However when examining (or copy out) the stored data, it is somewhat different than the original:
    original:
    <?xml version="1.0"?>
    <Airport>
    <Name>airport1.fly</Name>
    <AirportCode>SFO</AirportCode>
    <AirportName>San Francisco</AirportName>
    <FolderPath>.</FolderPath>
    </Airport>
    As recreated by the renderer:
    <?xml version = '1.0' standalone = 'yes'?>
    <AIRPORT>
    <Name>airport1.fly</Name>
    <Owner RefType="Name" ClassName="DIRECTORYUSER">system<!--ID # 96--></Owner>
    <ACL RefType="Name" ClassName="SYSTEMACCESSCONTROLLIST">Published<!--ID # 422--></ACL>
    <CreateDate format="MMM dd HH:mm">Nov 09 15:44</CreateDate>
    <Creator RefType="Name" ClassName="DIRECTORYUSER">system<!--ID # 96--></Creator>
    <LastModifyDate format="MMM dd HH:mm">Nov 09 15:44</LastModifyDate>
    <LastModifier RefType="Name" ClassName="DIRECTORYUSER">system<!--ID # 96--></LastModifier>
    <LockState>0</LockState>
    <AIRPORTCODE>SFO</AIRPORTCODE>
    <AIRPORTNAME>San Francisco</AIRPORTNAME>
    </AIRPORT>
    We'd like (to have cake and eat it) to be able to store parsed xml (to support queryability, indexing, reporting, etc.) but also want to be able to recreate the original xml UNCHANGED.
    If we don't parse it, we can have unchanged data, but then limit the queryability. Other than storing the xml document twice (both parsed and not) is there any way to do this?
    null

    I am glad that you solve this question and share the solution with us, I will mark your reply as the answer.
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • What is the difference between Topic Keywords and Index File Keywords?

    What is the difference between Topic Keywords and Index File Keywords? Any advantages to using one over the other? Do they appear differently in the generated index?
    RH9.0.2.271
    I'm using Webhelp

    Hi there
    When you create a RoboHelp project you end up with many different ancillary files that are used to store different bits of information. Many of these files bear the name you assigned to the project at the time you created it. The index file has the project name and it ends with a .HHK file extension. (HHK meaning HTML Help Keywords)
    Generally, unless you change RoboHelp's settings, you add keywords to this file and associate topics to the keywords via the Index pod. At the time you compile a CHM or generate other types of output, the file is consulted and the index is built.
    As I said earlier, the default is to add keywords to the Index file until you configure RoboHelp to add the keywords to the topics themselves. Once you change this, any keyword added will become a META tag in the topic code. If your keyword is BOFFO, the META tag would look like this:
    <meta name="MS-HKWD" content="BOFFO" />
    When the help is compiled or generated, the Index (.HHK) file is consulted as normal, but any topics containing keywords added in this manner are also added to the Index you end up with. From the appearance perspective, the end user woudn't know the difference or be able to tell. Heck, if all you ever did was interact with the Index pod, you, as an author wouldn't know either. Well, other than the fact that the icons appear differently.
    Operationally, keywords added to the topics themselves may hold an advantage in that if you were to import these topics into other projects, the Index keywords would already be present.
    Hopefully this helps... Rick

  • Different b/w index rebuild and index rebuild online

    hi..guys could u plz tel me difference between index rebuild and index rebuild online

    There is no difference in both the commands. Both will rebuild the index structure from the scratch.But in the first case with only Rebuild, as long as the index, its temporary segment is not prepared and merged together, index is not available for the other users for use. The Online clause makes the index available for others even while being rebuild.
    Rebuilding index online has the same concept of creating them online to some extent,
    http://download.oracle.com/docs/cd/B10501_01/server.920/a96521/indexes.htm#3062
    HTH
    Aman....

  • Parse and output XML document while preserving attribute order

    QUESTION: How can I take in an element with attributes from an XML and output the same element and attributes while preserving the order of those attributes?
    The following code will parse and XML document and generate (practically) unchanged output. However, all attributes are ordered a-z
    Example: The following element
    <work_item_type work_item_db_site="0000000000000000" work_item_db_id="0" work_item_type_code="3" user_tag_ident="Step" name="Work Step" gmt_last_updated="2008-12-31T18:00:00.000000000" last_upd_db_site="0000000000000000" last_upd_db_id="0" rstat_type_code="1">
    </work_item_type>is output as:
    <work_item_type gmt_last_updated="2008-12-31T18:00:00.000000000" last_upd_db_id="0" last_upd_db_site="0000000000000000" name="Work Step" rstat_type_code="1" user_tag_ident="Step" work_item_db_id="0" work_item_db_site="0000000000000000" work_item_type_code="3">
    </work_item_type>As you may notice, there is no difference in these besides order of the attributes!
    I am convened that the problem is not in the stylesheet.xslt but if you are not then it is posted bellow.
    Please, someone help me out with this! I have a feeling the solution is simple
    The following take the XML from source.xml and outputs it to DEST_filename with attributes in a-z order
    Code:
    private void OutputFile(String DEST_filename, String style_filename){
         //StreamSource stylesheet = new StreamSource(style_filename);
         try{
              File dest_file = new File(DEST_filename);
              if(!dest_file.exists())
                  dest_file.createNewFile();
              TransformerFactory tranFactory = TransformerFactory.newInstance();
              Transformer aTransformer = tranFactory.newTransformer();
              aTransformer.setOutputProperty(OutputKeys.ENCODING, "UTF-8");
              Source src = new DOMSource("source.xml");
              Result dest = new StreamResult(dest_file);
              aTransformer.transform(src, dest);
              System.out.println("Finished");
         catch(Exception e){
              System.err.print(e);
              System.exit(-1);
        }

    You can't. The reason is, the XML Recommendation explicitly says the order of attributes is not significant. Therefore conforming XML serializers won't treat it as if it were significant.
    If you have an environment where you think that the order of attributes is significant, your first step should be to reconsider. Possibly it isn't really significant and you are over-reaching in some way. Or possibly someone writing requirements is ignorant of this fact and the requirement can be discarded.
    Or possibly your output is being given to somebody else who has a defective parser which expects the attributes to be in a particular order. You could quote the XML Recommendation to those people but often XML bozos are resistant to change. If you're stuck writing for that parser then you'll have to apply some non-XML processing to your output to fix it up on their behalf.

  • DB02 view is empty on Table and Index analyses  DB2 9.7 after system copy

    Dear All,
                 I did the Quality refresh by System copy export/import method. ECC6 on HP-UX DB29.7.
    After Import Runstats status n Db02 for Table and Index analysis was empty and all value showing '-1'. Eventhough
    a) all standard backgrnd job scheduled in sm36
    b) Automatic runstats are enabled in db2 parameters
    c) Reorgchk all scheduled periodically from db13 and already ran twice.
    4) 'reorgchk update statistics on table all' was also ran on db2 level.
    but Run stats staus in db02 was not getting updated. Its empty.
    Please suggest.
    Regards
    Vinay

    Hi Deepak,
    Yes, that is possible (but only offline backup). But for the new features like reclaimable tablespace (to lower the high watermark)
    it's better to export/import with systemcopy.
    Also with systemcopy you can use index compression.
    After backup and restore you can have also reclaimable tablespace, but you have to create new tablespaces
    and then work with db6conv and online table move to move one tablespace online to the new one.
    Best regards,
    Joachim

  • Ceartion of User Defined Field in EXCHANGE RATE AND INDEXES

    Hi Experts,
                     I want to create  User Defined Field in EXCHANGE RATE AND INDEXES.But while creating the UDF from User Defined Field-Management unable to find the table for it.Write now My Client are using SAP B1 2007 Ptach-08.Is there any way out to create user defined field in EXCHANGE RATE AND INDEXES.
    Plz help me out on this issue.
    with regards,
    Pankaj K and Kamlesh N

    Pankaj,
    When you do the Manage User Fields area to define a UDF, all the possible areas where UDF's can be created in B1 is listed.  You would be able to create UDF's only on these.
    Suda

  • Table files and Index files 2GB on Windows 2003 Server SP2 32-bit

    I'm new to Oracle and I've ran into the problem where my Table files and Index files are > 2GB. I have an Oracle instance running version 10.2.0.3.0. I have a number of tables file and index files that have a current files size of 1.99GB. My Oracle crashes about three times a week because of a "Write Fault/Failure. I've detemined that the RDBM is trying to write a index or table files > 2GB. When this occurs it crashes.
    I've been reading the Oracle knowledge base that it suggest that there is a fix or release of Oracle 10g to resolve this problem. However, I've been unable to locate any fix or release to address my issue. Does such a fix or release exist? How do I address this issue? I'm from the world of MS SQL and IBM DB2 and we don't have this issue. I am running and NTFS files system. Could this be issue be related to an Windows Fix?
    Surely Oracle can handel databases > 2GB.
    Thanks in advance for any help.

    After reading your response it appears that my real problem has to do with checking pointing. I've included below a copy of the error message:
    Oracle process number: 8
    Windows thread id: 3768, image: ORACLE.EXE (CKPT)
    *** 2008-07-27 16:50:13.569
    *** SERVICE NAME:(SYS$BACKGROUND) 2008-07-27 16:50:13.569
    *** SESSION ID:(219.1) 2008-07-27 16:50:13.569
    ORA-00206: Message 206 not found; No message file for product=RDBMS, facility=ORA; arguments: [3] [1]
    ORA-00202: Message 202 not found; No message file for product=RDBMS, facility=ORA; arguments: [D:\ELLIPSE_DATABASE\CONTROL\CTRL1_ELLPROD1.CTL]
    ORA-27072: Message 27072 not found; No message file for product=RDBMS, facility=ORA
    OSD-04008: WriteFile() failure, unable to write to file
    O/S-Error: (OS 33) The process cannot access the file because another process has locked a portion of the file.
    error 221 detected in background process
    ORA-00221: Message 221 not found; No message file for product=RDBMS, facility=ORA
    ORA-00206: Message 206 not found; No message file for product=RDBMS, facility=ORA; arguments: [3] [1]
    ORA-00202: Message 202 not found; No message file for product=RDBMS, facility=ORA; arguments: [D:\ELLIPSE_DATABASE\CONTROL\CTRL1_ELLPROD1.CTL]
    ORA-27072: Message 27072 not found; No message file for product=RDBMS, facility=ORA
    OSD-04008: WriteFile() failure, unable to write to file
    O/S-Error: (OS 33) The process cannot access the file because another process has locked a portion of the file.
    Can you tell me why I'm having issues with checking point and the control file?
    Can I rebuild the control file if it s corrupt?
    The problem has been going on since April 2008. I'm takening over the system.
    Thanks

  • Best practice for PK and indexes?

    Dear All,
    What is the best practice for making Primary Key and indexes? Should we keep them in the same tablespace as table or should we create a seperate tableapce for all indexes and Primary Key? Please note I am talking about a table that has 21milion rows at the moment and increasing 10k to 20k rows daily. This table is also heavily involved in daily reports and causing slow performance. Currently the complete table with all associated objects such as indexes and PK is stored in one seperate tablespace. If my way is right then please advise me how can I improve the performance of retrival or DML operation on this table?
    Thanks in advance..
    Zia Shareef

    Well, thanks for valueable advices... I am using Oracle 8i and let me tell you exact problem...
    My billing database has two major tables having almost 21 millions rows each... one has collection data and other one for invoices... many reports are showing the data with the joining of Customer + Collection + Invoices tables.
    There are 5 common fields in between invoices(reading) and collection tables
    YEAR, MONTH, AREA_CODE, CONS_CODE, BILL_TYPE(adtl)
    My one of batch process has following update and it is VERY VERY SLOW:
    UPDATE reading r
    SET bamount (SELECT sum(camount)
    FROM collection cl
    WHERE r.ryear = cl.byear
    AND r.rmonth = cl.bmonth
    AND r.area_code = cl.area_code
    AND r.cons_code = cl.cons_code
    AND r.adtl = cl.adtl)
    WHERE area_code = 1
    tentatively area_code(1) is having 20,000 consumers
    each consuemr may have 72 invoices and against these invoices it may have 200 rows in collection tables (system have provision to record partial payment against one invoice)
    NOTE: Please note presently my process is based on cursors so the above query runs for one consumer at one time but just for giving an idea I have made it for whole area.
    Mr. Yingkuan, can you please tell me how can I check that the table' statistics is not current and how can I make it current. Is it really effect performance?

  • 'unable to connect' and  index.php

    Hi. 
    I am developing a Web Site and index.php is my point of entry.
    Document Root     Library/WebServer/Documents
    so my path is:       Library/WebServer/Documents/dwwdSite
    httpd.conf file is modified to add index.php  and have it listed first.
    <IfModule dir_module>
    DirectoryIndex   index.php   index.html
    <IfModule>
    Troubleshooting:
    I was using Netbeans IDE and when I ran index.php it opened in the browser.
    When I launched 'any' of my index.php files from Netbeans IDE, they opened correctly in the brower
    I am NOW using DreamweaverCC and when I run the index.php Error Message ' Unable to Connect'.
    For the last 2 days I have been working on this and I am completely stuck.
    This morning I thought of another way to test the 'unable to connect' error.
    I decided to copy this same file into Netbeans IDE and I NOW get the same Error Message ' Unable to Connect'
    when running index.php from Netbeans.
    Somehow, my settings are not correctly configured anymore.
    Here are my screenshots from Dreamweaver > manage sites.
    I believe that this is a rather simple fix that I am somehow not seeing.
    Maybe some can spot some mistake.
    I appreciate your help and explanation.

    Hi Sudarshan.
    You have been very kind and very clear in your explanation.
    One of the very best that I have ever communicated with on this forum !
    I have checked many, many things.
    I wanted to make certain that I killed apache and restarted it.
    I do not think it is RUNNING at all.
    1.
    myNameMacBookPro:~ myName$ pwd
    /Users/myName
    2.
    myNameMacBookPro:~ myName$ ps -ax | grep http
    1892 ttys000    0:00.00 grep http
    3.
    myNameMacBookPro:  myName$ hostname
    local
    4.
    myNameMacBookPro:etc myName$ cd apache2
    reginaMacBookPro:apache2 myName$ ls
    extra            httpd.conf.pre-update    mime.types        other
    httpd.conf        magic            original        users
    5.
    myNameMacBookPro:apache2 myName$ sudo nano httpd.conf
    myNameMacBookPro:apache2 myName$ sudo apachectl -k restart
    Syntax error on line 1 of /private/etc/apache2/users/myNameBU.conf:
    Invalid command '{\\rtf1\\ansi\\ansicpg1252\\cocoartf1187\\cocoasubrtf370', perhaps misspelled or defined by a module not included in the server configuration
    httpd not running, trying to start
    myNameMacBookPro:apache2 myName$
    6.
    The above code may be a hint at part of the problem.
    I created myNameBU.conf   as a backup when I was editing the file.
    QUESTION.  Why is /private/etc/apache2/users/myNameBU.conf:
    being referenced above ?
    7.
    I scanned my ports from preferences and found these two which I believe are what you stated they should be.
    port 80 is html
    port 8080 is html-alt
    8.
    I have modified this line and tried it both ways restarting apachectl each time.
    AllowOverride None
    AllowOverride All
    9.
    And here is a part of my httpd.conf.
    excerpts.
    httpd.conf
    # User/Group: The name (or #number) of the user/group to run httpd as.
    # It is usually good practice to create a dedicated user and group for
    # running httpd, as with most system services.
    User _www
    Group _www
    # DocumentRoot: The directory out of which you will serve your
    # documents. By default, all requests are taken from this directory, but
    # symbolic links and aliases may be used to point to other locations.
    DocumentRoot "/Library/WebServer/Documents"
    # Each directory to which Apache has access can be configured with respect
    # to which services and features are allowed and/or disabled in that
    # directory (and its subdirectories).
    # First, we configure the "default" to be a very restrictive set of
    # features.
    <Directory />
        Options FollowSymLinks
        AllowOverride None
        Order deny,allow
        Deny from all
    </Directory>
    10.
    I took a look at this too.
    usr/bin
    #path to httpd binary including options if necessary
    HTTPD = "usr/sgin/httpd"
    # pick up any environmental variables if test - f /usr/sbin/envvars; then ./usr/sbin/envvars
    fi
    STATUSURL = "http://localhost:80/server-status
    11.
    Just a reminder.
    I started writing .php scripts with Netbeans IDE.
    The programs ran:  I got output in the brower.  Things worked just fine !
    I started writing .php scripts with DreamweaverCC.
    The programs NEVER ran.
    I have always gotten 'Unable to Connect"
    Firefox/Safari can't establish a connection to the server at localhost.'
    12.
    QUESTION.
    At one point I was on the phone with a member of the Adobe Technical Support Team.
    They connect to my desktop (only) remotely.  I am very cautious about this.
    Could something have been inadvertently changed when they did this ?
    They do need to connect through a PORT - yes ?
    This is very frustrating and I am loosing days of work.
    I want to get back to Web Development
    I love Adobe products but this should not be such a huge obstacle.
    When I called Technical support (after much troubleshooting individually and on this forum)
    the fellow told me that the ONLY support FTP - not LOCALHOST.
    This makes no sense.  People develop 'locally' then put it ftp it in the production stage.
    Again, I appreciate your assistance.

  • 'unable to connect' and 'localhost' and index.php and dreamweaverCC

    Hi. 
    I am developing a Web Site and index.php is my point of entry.
    Document Root Library/WebServer/Documents
    so my path is: Library/WebServer/Documents/dwwdSite
    httpd.conf file is modified to add index.php  and have it listed first.
    <IfModule dir_module>
    DirectoryIndex   index.php   index.html
    <IfModule>
    Troubleshooting:
    I was using Netbeans IDE and when I ran index.php it opened in the browser.
    When I launched 'any' of my index.php files from Netbeans IDE, they opened correctly in the brower
    I am now using DreamweaverCC and when I run the index.php Error Message ' Unable to Connect'.
    For the last 2 days I have been working on this and I am completely stuck.
    This morning I thought of another way to test the 'unable to connect' error.
    I decided to copy this same file into Netbeans IDE and I NOW get the same Error Message ' Unable to Connect'
    when running index.php from Netbeans.
    Somehow, my settings are not correctly configured anymore.
    Here are my screenshots from Dreamweaver > manage sites.
    I believe that this is a rather simple fix that I am somehow not seeing.
    Maybe some can spot some mistake.
    I appreciate your help and explanation.

    Site window settings.
    Site Name: dwwdSite
    Local site folder: /Library/WebServer/Documents/dwwdSite
    Server window settings.
    Server Name: testing Server
    Address: Macintosh HD/Library/WebServer/Documents/dwwdSite
    Connect using: Local/network
    Testing: yes (checked)
    Server folder: /Library/WebServer/Documents/dwwdSite
    (I also tried this: Server folder: /Library/WebServer/Documents)
    Web URL: http://www.localhost/dwwdSite
    Server Advanced tab: (within server window settings)
    Testing server: PHP MySQL
    Advanced Settings window.
    Local info: Web URL: http://www.localhost/dwwdSite
    Enable cache: yes (checked)

  • Conforming and Indexing Errors, Media Pending, Audio won't play in timeline

    I'm working on a desktop PC which is running Windows 7 Professional 64-bit and Adobe Premiere Pro (version CS5.5). It's currently utilizing a second gen. 3.4Ghz i7 2600 processor, 16GB of 1600Mhz RAM, 64GB solid-state drive and a ASUS P8Z68-V Intel Z68 Motherboard with onboard audio (Realtek ALC892 chipset) and onboard video. My problem is this:
    The conforming and indexing of all of my imported media never seems to finish regardless of how many times I reopen the project file and wait for it. On the lower right-hand portion of the screen, next to the conforming/indexing progress bar, is a little red "X". When clicked, it pops up with a list of errors that read: "An unexpected error occurred while performing a conform action on the following file...". As a result, my audio channels have no waveform and during playback there are no audible tones or levels. On some video clips there's just text that reads "Media Pending". This only appears to happen with project files that I saved on external hard drives, and I suspect it has something to do with the Media Cache Files folder and how Premiere Pro locates these conform/index files. I've also encountered this problem in CS3 and CS4.
    I have a few questions:
    1) How do I avoid error messages in regards to indexing and conforming
    2) How do you know when indexing/conforming has completed itself? (there doesn't seem to be a progress log or a list of commands/executions)
    3) Indexing and conforming appears to be an automatic process, but is there a way to do it manually?
    4) What's the best way to setup your media cache files when you click EDIT > PREFERENCES > MEDIA?
    5) If I have approximately 1 hour of footage, what's an average wait time for conforming/indexing? What about 5 hours of footage? 10?
    6) Adobe recommends not editing until the conforming and indexing has completed itself-- how important is this?
    7) Sometimes it appears as though the conforming and indexing has finished, but then I still have problems with playback. Do I have to reopen the project for it to continue with the conforming/indexing progress? I've already determined that the video file I'm working with is intact and free of any corruption.
    I'm fine with having to wait for a project to conform and index, but it never seems to complete itself! Any help regarding this matter would be greatly appreciated.

    Harm filled in pretty much all the salient details, but I'll do another pass here.
    1) How do I avoid error messages in regards to indexing and conforming
    Two parts here.  One, conforming only happens for certain media files, ie the ones where performance is critical and we can't depend on extracting the audio fast enough for realtime playback.  That's basically anything in an .mpeg wrapper, or AVCHD material.  So if you edit XDCAM HD/EX or P2, or RED, or even AVIs or QT, those formats don't require audio conforming.
    If you're stuck editing AVCHD or MPEG2, then it needs to conform.  But, that being said, you shouldn't be getting errors in the first place. I think it's related to your external drives.  More below...
    2) How do you know when indexing/conforming has completed itself? (there doesn't seem to be a progress log or a list of commands/executions)
    Nope, you have a progress status bar indicating which file it's working on.  If there's an error, it shows up in the events panel.
    3) Indexing and conforming appears to be an automatic process, but is there a way to do it manually?
    No.
    4) What's the best way to setup your media cache files when you click EDIT > PREFERENCES > MEDIA?
    While some people like having the check box for having the conform files beside the media, I hate it.  Yes, it means that if you move the project to a different system & reopen, it means that you potentially can avoid recreating CFA files, but I find the drive littering not worth it.  I much prefer having setting the Media prefs to point to a specific media drive.  Usually a raid, if available.  Definitely not an external drive that you disconnect & walk away with.  If you don't have a permanent raid on your system, then preferably a dedicated internal drive for media (think along the lines as your Photoshop 'scratch disk').  Failing that, leave it on your C: drive, although with a 64 Gig SSD, you probably don't have much room for transient temporaries.
    5) If I have approximately 1 hour of footage, what's an average wait time for conforming/indexing? What about 5 hours of footage? 10?
    Like Harm said.  Totally dependant on the media container & the speed of your drive i/o.  The conforming is iterating through the entire file & pulling audio data, so it's not CPU intensive, it's all i/o.
    6) Adobe recommends not editing until the conforming and indexing has completed itself-- how important is this?
    If you're trying to play/scrub while conforming, it's going to be pokey.  Esp. if you're trying to access the file that's actively being conformd.  As I just said, we're hitting the files for all the audio.  The i/o is being saturated already, so unless you have a stellar raid, you don't have much headroom.
    7) Sometimes it appears as though the conforming and indexing has finished, but then I still have problems with playback. Do I have to reopen the project for it to continue with the conforming/indexing progress? I've already determined that the video file I'm working with is intact and free of any corruption.
    You should be good to go.  Sounds like there's something else at play here.
    Okay, back to what I think is wrong:  you don't mention what kind of external drives you're using.  You're making a bad assumption that blowing away conformed files & doing a reconform is buggy - I doubt it, as that's the same process that happened when you initially brought in the files.  I've blown away my media cache folder multiple times and have never seen failures on reconform.  So it's got to be one of two things:  either a read error from the source when attempting to pull the audio, or a write error to the destination.  Now I don't know where you currently are pointing the media cache directory, or what your source drive is, so I can only speculate.
    My suggestion is to do some elimination.   Copy one of the files that failed on you to your C drive, & target your media cache directory also to C:.  Pick a new project, import your copied file, confirm that it conforms correctly & behaves.   Then, try to use the same clip from your external drive, keeping the media cache to C:.  If that's still good, then try targeting another (local/internal) drive as your media cache target; close/restart, then import the clip from C:, and then import the clip from your external drive.  This troubleshooting should give us something.
    PS, if you're trying to edit from external USB drives, good luck.  I find it a major PITA that I avoid as much as possible.  Firewire isn't much better.  I know some people do it successfully, but I think it's a road fraught with peril.  These devices are generally not designed for heavy duty I/O and a flaky connection or drive is nothing but pain.
    Cheers

  • Report to find all table and index sizes

    Hi all,
    Good day..
    Is there any report.sql or so to find out the sizes of all the tables and indexes in a database.
    thanks,
    baskar.l

    1.To get table size
    What will be the table size if?
    <or>
    break on report
    set line 200
    COMPUTE SUM LABEL "Total Reclaimable Space" OF "KB Free Space" ON REPORT
    column "Table Size" Format a20
    column "Actual Data Size" Format a20
    column "KB Free Space" Format "9,99,999.99"
    select table_name,
    round((blocks*8),2)||'kb' "Table size",
    round((num_rows*avg_row_len/1024),2)||'kb' "Actual Data size",
    pct_free,
    round((blocks*8),2) - (round((blocks*8),2)*pct_free/100) - (round((num_rows*avg_row_len/1024),2)) "KB Free Space"
    from user_tables
    where round((blocks*8),2) - (round((blocks*8),2)*pct_free/100) - (round((num_rows*avg_row_len/1024),2)) > 0
    order by round((blocks*8),2) - (round((blocks*8),2)*pct_free/100) - (round((num_rows*avg_row_len/1024),2)) desc
    2.To get index size
    How to size the Index
    Hth
    Girish Sharma

Maybe you are looking for

  • Can't import P2 Footage after 5.1.2 update

    This one has me stumped. After updating to 5.1.2 I can not import P2 footage. When I click on the import Panasonic P2 tab under the import menu, nothing happens. Then, I click on it one more time and the import window loads but there is nothing there

  • Delete every payroll results in the system

    Hi experts, There is a special Test system where External useres can login to the system, so we'd like to delete every paroll results in the system. My question is how can I do that * correctly & efficiently*? I know that there is a report for that: 

  • Save PDF to folder as JPG not working....

    I can't figure out why all of a sudden Aperture book pages exported as JPG's aren't showing up anywhere. I did change the dpi setting to 300 and have successfully exported other books recently, but today I finished a book layout and wanted to export

  • Has anyone tried the "Table Filters" with Numbers 3.0 (Maverick).

    I use the formula If(isblank(A5),"","Data"). the previous version of Numbers treated the "" as a blank cell and the Table Filters did not display that row if I had the "Reorginize" filter set to "Is Not Blank". The Maverick version treats the "" as t

  • FCPX error message 1028 error 1

    Hi there I have recently upgraded FCPX to 10.0.09.  When I export my video to the master file I am now receiving an error message, which says: The share operation Master File has failed.  The operation could not be completed because an error occured