Corrupted database the reason the index

Hi there 
Due to sudden power outage I got my AD database corrupted, I am running DC on win server 2012 
and this is what contained in the log file 
Log Name:      Directory Service
Source:        NTDS ISAM
Date:          7/9/2014 4:12:01 PM
Event ID:      467
Task Category: Database Corruption
Level:         Error
Keywords:      Classic
User:          N/A
Computer:      DC
Description:
NTDS (556) NTDSA: Database C:\Windows\NTDS\ntds.dit: Index INDEX_00020078 of table datatable is corrupted (0).
Event Xml:
<Event xmlns="http://schemas.microsoft.com/win/2004/08/events/event">
  <System>
    <Provider Name="NTDS ISAM" />
    <EventID Qualifiers="0">467</EventID>
    <Level>2</Level>
    <Task>12</Task>
    <Keywords>0x80000000000000</Keywords>
    <TimeCreated SystemTime="2014-07-09T23:12:01.000000000Z" />
    <EventRecordID>3204</EventRecordID>
    <Channel>Directory Service</Channel>
    <Computer>DC.training.local</Computer>
    <Security />
  </System>
  <EventData>
    <Data>NTDS</Data>
    <Data>556</Data>
    <Data>NTDSA: </Data>
    <Data>INDEX_00020078</Data>
    <Data>datatable</Data>
    <Data>C:\Windows\NTDS\ntds.dit</Data>
    <Data>0</Data>
  </EventData>
</Event>

That is the index over the UsnChanged attribute - replication will stop working as this is used by the DRA to perform replications.
If you have other DCs in this domain, I would recommend to power off this machine and discard it and perform metadata cleanup on another DC:
http://support.microsoft.com/kb/216498/sv
To fix this issue you have to perform an offline defrag:
http://technet.microsoft.com/en-us/library/cc794920(v=ws.10).aspx
Enfo Zipper
Christoffer Andersson – Principal Advisor
http://blogs.chrisse.se - Directory Services Blog

Similar Messages

  • Errors in the OLAP storage engine: An error occurred while processing the indexes for the partition of the measure group of the cube from the database.

    Errors in the OLAP storage engine: An error occurred while processing the indexes for the partition of the measure group of the cube from the database.
    I have dropped and recreated a fact table to refresh the data as the table is having identity columns. So not able to directly insert data. After that cube is throwing above error.
    Please suggest.

    Hi Md,
    It hard to analyse the issue base on the limited information, Are there any error message on the log? Under C:\Program Files\Microsoft SQL Server\MSAS11.\OLAP\Log (this could be different for your server), there are 3 log files that are generated:
    msmdrrv.log
    FlightRecorderCurrent.trc
    FightRecorderBack.trc
    The msmdrrv.log simply points to the other two logs. There might be some detail information for this error. Please provide us the detail information about it, so that we can make further analysis.
    Regards,
    Charlie Liao
    TechNet Community Support

  • Corrupted database for download in the link

    I have Downloaded Oracle 8i Database from the Link
    http://download-east.oracle.com/otn/linux/oracle8i/linux81701.tar
    I found out that the following file is corrupted in the server as i tried downloaded twice.

    Not sure why you are not seeing it. I just checked the App Store on my iPhone and in iTunes on my Mac, and it shows up in both places, indicating that it has updates for iOS 7 compability.....(and is an iPhone/iPad app)....
    GB

  • Can we drop and then recreate the indexes when database is open.

    Can we drop and then recreate the indexes when database is open.

    I am trying to drop the index not happening..any suggestion..
    I tried :
    drop index index_name >>> index does not exist, though it exist on particular table
    drop index table_name.index_name
    drop index index_name on table_name; >> SQL command not properly ended
    Could you suggest me how would i drop it

  • What is the source of this Crawl Error `The item could not be indexed successfully because the item failed in the indexing subsystem`

    Once in a while my full-crawl fails and stops working properly. The crawllogs show this error for all items crawled.
    The item could not be indexed successfully because the item failed in the indexing subsystem. ( The item could not be indexed successfully because the item failed in the indexing subsystem.; Caught exception when preparing generation GID[7570]: (Previous generation
    (last=GID[7569], curr=GID[7570]) is still active. Cannot open GID[7570]); Aborting insert of item in Link Database because it was not inserted to the Search Index.; ; SearchID = F201681E-AF1B-45D2-BFFD-6A2582D10C19 )
    The full crawls starts out ok, after a while (1.5 hours into the process, 50% of all the data) suddenly no more items can be added to the index.
    The index seems to be stuck. The index files on disk are no longer updated (Located in D:\Microsoft
    Office Servers\15.0\Data\Office Server\Applications\Search\Nodes\BAADC4\IndexComponent3\storage\data\SP4d91e6081ac3.3.I.0.0\ms\%default) The Index and Admin component start to report these error in the ULS logs:
    NodeRunnerIndex: Journal[SP4d91e6081ac3]:
    Rolling back GID[7570] to GID[7569] prepGen=GID[7569]
    NodeRunnerIndex: Remote
    service invocation: method=RollbackGeneration() Service = {  Implementation type=Microsoft.Ceres.SearchCore.ContentTargets.IndexRouter.IndexRouter  Component: SP4d91e6081ac3I.0.0.IndexRouter  Exposer Name: GenerationContentTarget} terminated
    with exception: System.InvalidOperationException: Illegal state transition in SP4d91e6081ac3I.0.0.FastServer.FSIndex: Rollback -> Rollback
    NodeRummerAdmin: RetryableInvocation[SP4d91e6081ac3]:
    Exception invoking index cell I.0.0. Retrying in 16 seconds: System.InvalidOperationException: Illegal state transition in SP4d91e6081ac3I.0.0.FastServer.FSIndex: Rollback -> Rollback
    Looks to me the index has troubles updating/merging 'generations'. But the exact working of the indexer is not documented (as far is I know). Let alone how to fix this.
    Other (maybe related) observations 
    Just before the errors start the NodeRunnerIndex starts a 'checkpoint': Journal[SP4d91e6081ac3]:
    Starting checkpoint because forceCheckpoint is true.which ends a few moments later with Journal[SP4d91e6081ac3]:
    All journal users have completed checkpoint Checkpoint[7560-7569].
    Also just before the errors start to appear a TimerJob starts: Name=Timer
    Job job-application-server-admin-service. This timerjobs does some strange things to the search topology: Synchronizing
    Search Topology for application 'Search Service Application' with active topology [...] and Activating
    components. Previous topology:   ---  New Topology: TopologyId: [...] followed by Starting
    to execute Index RedistributeData method.
    And right after these two evente the errors start to occur. (each row is a ULS log enrty)
    INFO : fsplugin: IndexComponent3-bd83a8aa-923b-4526-97e8-47eac0986ff7-SP4d91e6081ac3.I.0.0 (4236): Prepare generation: 324 documents
    IndexRouter[SP4d91e6081ac3]: Caught exception when preparing generation GID[7570]: (External component has thrown an exception.): System.Runtime.InteropServices.SEHException (0x80004005): External component has thrown an exception.
    GenerationDispatcher[SP4d91e6081ac3]: Failed to prepare GID[7570] in 453 ms, failed on cells: [I.0.0], stale services: []
    Journal[SP4d91e6081ac3]: Rolling back GID[7570] to GID[7569] prepGen=GID[7569]
    Remote service invocation: method=RollbackGeneration() Service = { Implementation type=Microsoft.Ceres.SearchCore.ContentTargets.IndexRouter.IndexRouter Component: SP4d91e6081ac3I.0.0.IndexRouter Exposer Name: GenerationContentTarget} terminated with exception: System.Runtime.InteropServices.SEHException (0x80004005): External component has thrown an exception.
    RetryableInvocation[SP4d91e6081ac3]: Exception invoking index cell I.0.0. Retrying in 2 seconds: System.Runtime.InteropServices.SEHException (0x80004005): External component has thrown an exception.
    Journal[SP4d91e6081ac3]: Rolling back GID[7570] to GID[7569] prepGen=GID[7569]
    The Question What is causing this? And how to prevent it? It happend twice in two weeks
    nows. Out of the blue, no config change has been made, all disks have enough space.
    Known Fix This resolves the problem, but doesn's address the root-cause of the problem!
    Stop all crawls.
    Wait a few minutes to let the crawl come to a complete stop.
    Reset the index (clearing all!)
    Start a full-crawl. In the meantime no search is available to the end user (boohoo!)

    Hi,
    I searched for the similar error log, the issue is finally solved by adding more drive space even though they though there is plenty space already.
    https://social.technet.microsoft.com/Forums/office/en-US/d06c9b2c-0bc1-44c6-b83a-2dc0b70936c4/the-item-could-not-be-indexed-successfully-because-the-item-failed-in-the-indexing-subsystem?forum=sharepointsearch
    http://community.spiceworks.com/topic/480369-the-item-could-not-be-indexed-successfully
    From your decription, the issue seems to occur to your full crawl. There is one point in best practise for crawling:
    Run full crawls only when it is necessary, the reasons to do a full crawl are as below:
    https://technet.microsoft.com/en-us/library/jj219577.aspx#Plan_full_crawl
    https://technet.microsoft.com/en-us/library/dn535606.aspx
    Regards,
    Rebecca Tu
    TechNet Community Support
    Please remember to mark the replies as answers if they help, and unmark the answers if they provide no help. If you have feedback for TechNet Support, contact
    [email protected]

  • How to transfer the indexes from one DB To another DB

    Hi All,
    I have one question.I have to transfer all the database object from one database to another. and I have transfer it.Now I have to create an index on that tables then how can i know on which columns that indexes were created.and How I will create it in new database. Or Is there another way to transfer that indexes from old database to new database.
    Thanx in advance.

    user13310428 wrote:
    How I will create it in new database. Or Is there another way to transfer that indexes from old database to new database.You have to use the DDL (create index/create table with index constraints/etc) commands to create indexes.
    You cannot "+transfer+" the index data like you can table data. The reason is that indexes contains the physical address of the rows indexed. These addresses are different on the new database - and thus make the index data on the old database useless. Instead, you need to re-create the indexes so that the new row addresses can be determined and used by the index.

  • Block Corruption in the SYSAUX tablespace

    Problem Description: Block Corruption in the SYSAUX tablespace :
    ==================================
    We are facing block corruption issue in the SYSAUX tablespace during rman backup..
    So currently we are using SET MAXCORRUPT FOR DATAFILE 3 TO 2; parameter to take the backup and it is running fine.. Now we have to fix the issue..
    Please let us know the steps to fix the issue..
    Please find the details below.
    Dbv output:
    sesoxpro33.p51x1> dbv file=/oradata/p51x/data/p51x_sysaux_01.dbf blocksize=32768 end=64000
    DBVERIFY: Release 10.2.0.3.0 - Production on Fri Aug 26 20:47:29 2011
    Copyright (c) 1982, 2005, Oracle. All rights reserved.
    DBVERIFY - Verification starting : FILE = /oradata/p51x/data/p51x_sysaux_01.dbf
    Page 64 is influx - most likely media corrupt
    Corrupt block relative dba: 0x00c00040 (file 3, block 64)
    Fractured block found during dbv:
    Data in bad block:
    type: 0 format: 0 rdba: 0x00000000
    last change scn: 0x0000.00000000 seq: 0x0 flg: 0x00
    spare1: 0x0 spare2: 0x0 spare3: 0x0
    consistency value in tail: 0x00000001
    check value in block header: 0x0
    block checksum disabled
    DBVERIFY - Verification complete
    Total Pages Examined : 64000
    Total Pages Processed (Data) : 22547
    Total Pages Failing (Data) : 0
    Total Pages Processed (Index): 23859
    Total Pages Failing (Index): 0
    Total Pages Processed (Other): 14586
    Total Pages Processed (Seg) : 0
    Total Pages Failing (Seg) : 0
    Total Pages Empty : 3007
    Total Pages Marked Corrupt : 1
    Total Pages Influx : 1
    Highest block SCN : 281617057 (2054.281617057)
    SELECT tablespace_name, segment_type, owner, segment_name FROM dba_extents
    WHERE file_id = 3 and 64 between block_id AND block_id + blocks - 1;
    TABLESPACE_NAME SEGMENT_TYPE OWNER
    SEGMENT_NAME
    SYSAUX SYS_LOB0000000594C00007$$
    OS & database Details :
    ==========
    Oracle Solaris on x86-64 (64-bit)
    Oracle Server - Enterprise Edition -- 10.2.0.3
    If any more informatin is needed, Please let me know..
    Thanks & Regards,
    Suresh Bommalata.

    refer:
    logical corruption found in the sysaux tablespace
    Check the metalink note: Identify the corruption extension using RMAN/DBV/ANALYZE etc - 836658.1
    Note 28814.1 - Handling Oracle Block Corruptions in Oracle7/8/8i/9i/10g
    Note 472231.1 - How to identify all the Corrupted Objects in the Database reported by RMAN
    http://download.oracle.com/docs/cd/B12037_01/server.101/b10734/rcmrecov.htm
    http://oracleinstance.blogspot.com/2010/05/block-recovery-using-rman-backup.html

  • Can I select the data of the indexes?

    Hi everybody,
    Oracle storages indexes in index type segments as I know.
    Is it possible to see the data of these index segments like a table or view?
    So I would like to see not only the structue of it with:
    select * from dba_segments SG where SG.segment_name = ’index_name’
    I think something like this:
    For example if I have a table named Table1.
    (please don’t see the format of the rowid):
    Rowid....field1
    1_1.......AA
    1_2.......AB
    1_3.......BB
    and I have an index on field1 named ndx_t1_f1.
    Create index ndx_t1_f1 on Table1(field1);
    I would like to see something like this
    (if it is a simple Btree index like the ndx_t1_f1):
    Select * from some_schema.???ndx_t1_f1???
    Rowid..........ParentRowid..........ValuePrefix..........SolutionRowid
    2_1
    2_11............2_1.......................A
    2_12............2_1.......................B
    2_111..........2_11.....................AA......................1_1
    2_112..........2_11.....................AB......................1_2
    2_121..........2_12.....................BB......................1_3
    And if i run:
    select field1 from table1 where filed1 like 'B%'
    oracle can use ndx_t1_f1 index like this:
    2_1 -> 2_12 -> 2_121 -> BB
    Yes, I know, there are some promlems with this imagine, but is it possible, or is it a silly question?
    Thanx: lados.

    Hi,
    You asked: „Why do you want to do that?”
    Well, I have a table T with fields F1, F2, F3 ...
    If I run:
    select F2, T.* from T where F1=12345;
    the result is one record and the value of F2 is ‘A’.
    But if I run:
    select F2, F1 from T where F1=12345;
    the result is one record and the value of F2 is ‘B’.
    And this is the same record. Oppps!!!!!
    I think the cause of the problem is the next:
    I have an index on F1, F2 named ndx_T_F1_F2
    In the second case: oracle uses only the index blocks and no table blocks, because all asked data exists in the index blocks.
    In the first case: oracle uses the table blocks too, because I asked all field.
    And I think there is a corrupted data on the index.
    I created a new index on (F1,F2,F3), and I suggest to use this.
    select /*+index(T, ndx_T_F1_F2_F3)*/ T.F2 from T where F1=12345;
    the result is one record and the value of F2 is ‘A’.
    But what is strange, oracle didn’t realise this inconsistent situation.
    (It could be an earlier crash, I don’t know, I’m new at this workplace)
    I think, the problem can solve with recreating the index ndx_T_F1_F2.
    But it’s a very special situation for me, and I would like to see the corrupted data concretly with my eyes.
    I think, that
    Select * from some_schema.???ndx_t1_f1???
    would I see like a view, but it would based on procedures and not tables.
    It would be enough to me, because I only could see the data in the index.
    But it is impossible as you wrote, isn’t it?
    Thanx: lados.

  • The plan doesn't use the index but the cost of INDEX FULL SCAN looks better

    Hi,
    Well, I'm sure I miss the boat... and if the question is pretty tricky, the answer is probably :"You're stupid Greg!". Well anyway, you'll probably be interested in trying to answer it as I've spent some times on it without any result ! I use Oracle XE on Windows...
    1) Below is my query and its plan. You'll find the full schema generation script at the end of this email. Look at the cost (468) of the plan and the cost of the same query when you force the use of the index (116). Why is this ?
    select count(distinct col5)
      2    from demo
      3      where col1 between 1 and 50000
      4        and col2=col1
      5        and col3=col1
      6        and col4=col1;
    Plan d'execution
    Plan hash value: 2072716547
    | Id  | Operation          | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT   |      |     1 |   116 |   468   (2)| 00:00:06 |
    |   1 |  SORT GROUP BY     |      |     1 |   116 |            |          |
    |*  2 |   TABLE ACCESS FULL| DEMO |     1 |   116 |   468   (2)| 00:00:06 |
    Predicate Information (identified by operation id):
       2 - filter("COL2"="COL1" AND "COL3"="COL1" AND "COL4"="COL1" AND
                  "COL1"<=50000 AND "COL2"<=50000 AND "COL3"<=50000 AND "COL4"<=50000 AND
                  "COL1">=1 AND "COL2">=1 AND "COL3">=1 AND "COL4">=1)2) When I force the use of an index (with a Hint), You'll see the cost of the plan is 116 which is definitly better than the TABLE ACCESS FULL (468) :
    SQL> l
      1  select /*+ index(demo demo_idx)*/ count(distinct col5)
      2    from demo
      3      where col1 between 1 and 50000
      4        and col2=col1
      5        and col3=col1
      6*       and col4=col1
    SQL> /
    Plan d'execution
    Plan hash value: 189561699
    | Id  | Operation                    | Name     | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT             |          |     1 |   116 |   437   (2)| 00:00:06 |
    |   1 |  SORT GROUP BY               |          |     1 |   116 |            |          |
    |   2 |   TABLE ACCESS BY INDEX ROWID| DEMO     |     1 |   116 |   437   (2)| 00:00:06 |
    |*  3 |    INDEX FULL SCAN           | DEMO_IDX |     1 |       |   436   (2)| 00:00:06 |
    Predicate Information (identified by operation id):
       3 - filter("COL2"="COL1" AND "COL3"="COL1" AND "COL4"="COL1" AND
                  "COL1"<=50000 AND "COL2"<=50000 AND "COL3"<=50000 AND "COL4"<=50000 AND
                  "COL1">=1 AND "COL2">=1 AND "COL3">=1 AND "COL4">=1)3) My question is why is plan1 used while plan2 should be considered better by the optimizer regarding the cost (to make the case even more complex, plan1 is actually more efficient but this is out of the scope of my question. I know that and I know why !).
    You'll find a script to generate the structures and data below. I can send you the 10053 traces if you what to go furthermore. Take care the index is a REVERSE index (Don't know if query rewrite should be enabled in order to take advantage of this type of index but it is set to "true" (and "trusted") :
    drop table demo;
    create table demo (col1 number not null,
        col2 number,
        col3 number,
        col4 number,
        col5 varchar2(500));
    begin
      for i in 1..100000 loop
        insert into demo values (i,i,i,i,'This column is used to raise the High Water Mark and '||
                                 ' the cost of an TABLE ACCESS FULL operation');
      end loop;
    end;
    commit;
    create index demo_idx on demo(col1,col2,col3,col4) reverse;
    exec dbms_stats.gather_table_stats(USER, 'DEMO', cascade=>true, -
      method_opt=>'FOR ALL COLUMNS SIZE 254', no_invalidate=>false) Any comments are welcome ! Best Regards,
    Gregory
    Message was edited by:
    arkzoyd... I've added the "pre" tags

    I suspect this has something to do with db_file_multiblock_read_count
    After running provided creation statements by you I got following results:
    SQL> show parameter multiblock
    NAME                                 TYPE        VALUE
    db_file_multiblock_read_count        integer     16
    SQL> set autotrace on
    SQL> select count(distinct col5)
      2   from demo
      3   where col1 between 1 and 50000
      4   and col2=col1
      5   and col3=col1
      6   and col4=col1
      7  /
    COUNT(DISTINCTCOL5)
                      1
    Execution Plan
    Plan hash value: 2072716547
    | Id  | Operation          | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT   |      |     1 |   116 |   375   (1)| 00:00:05 |
    |   1 |  SORT GROUP BY     |      |     1 |   116 |            |          |
    |*  2 |   TABLE ACCESS FULL| DEMO |     1 |   116 |   375   (1)| 00:00:05 |
    Predicate Information (identified by operation id):
       2 - filter("COL2"="COL1" AND "COL3"="COL1" AND "COL4"="COL1" AND
                  "COL1"<=50000 AND "COL2"<=50000 AND "COL3"<=50000 AND "COL4"<=5000
    0 AND
                  "COL1">=1 AND "COL2">=1 AND "COL3">=1 AND "COL4">=1)
    Statistics
            196  recursive calls
              0  db block gets
           1734  consistent gets
            850  physical reads
              0  redo size
            422  bytes sent via SQL*Net to client
            385  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              7  sorts (memory)
              0  sorts (disk)
              1  rows processed
    SQL> select /*+ index(demo demo_idx)*/ count(distinct col5)
      2   from demo
      3   where col1 between 1 and 50000
      4   and col2=col1
      5   and col3=col1
      6   and col4=col1
      7  /
    COUNT(DISTINCTCOL5)
                      1
    Execution Plan
    Plan hash value: 189561699
    | Id  | Operation                    | Name     | Rows  | Bytes | Cost (%CPU)| T
    ime     |
    |   0 | SELECT STATEMENT             |          |     1 |   116 |   431   (1)| 0
    0:00:06 |
    |   1 |  SORT GROUP BY               |          |     1 |   116 |            |
            |
    |   2 |   TABLE ACCESS BY INDEX ROWID| DEMO     |     1 |   116 |   431   (1)| 0
    0:00:06 |
    |*  3 |    INDEX FULL SCAN           | DEMO_IDX |     1 |       |   430   (1)| 0
    0:00:06 |
    Predicate Information (identified by operation id):
       3 - filter("COL2"="COL1" AND "COL3"="COL1" AND "COL4"="COL1" AND
                  "COL1"<=50000 AND "COL2"<=50000 AND "COL3"<=50000 AND "COL4"<=5000
    0 AND
                  "COL1">=1 AND "COL2">=1 AND "COL3">=1 AND "COL4">=1)
    Statistics
              1  recursive calls
              0  db block gets
          50426  consistent gets
            428  physical reads
              0  redo size
            422  bytes sent via SQL*Net to client
            385  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              1  sorts (memory)
              0  sorts (disk)
              1  rows processedNow I modify multiblock_read_count and full scan cost is going up although anyway Oracle by default chooses full scan instead of index access.
    SQL> alter session set db_file_multiblock_read_count = 8;
    Session altered.
    SQL> select count(distinct col5)
      2   from demo
      3   where col1 between 1 and 50000
      4   and col2=col1
      5   and col3=col1
      6   and col4=col1
      7  /
    COUNT(DISTINCTCOL5)
                      1
    Execution Plan
    Plan hash value: 2072716547
    | Id  | Operation          | Name | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | SELECT STATEMENT   |      |     1 |   116 |   463   (1)| 00:00:06 |
    |   1 |  SORT GROUP BY     |      |     1 |   116 |            |          |
    |*  2 |   TABLE ACCESS FULL| DEMO |     1 |   116 |   463   (1)| 00:00:06 |
    Predicate Information (identified by operation id):
       2 - filter("COL2"="COL1" AND "COL3"="COL1" AND "COL4"="COL1" AND
                  "COL1"<=50000 AND "COL2"<=50000 AND "COL3"<=50000 AND "COL4"<=5000
    0 AND
                  "COL1">=1 AND "COL2">=1 AND "COL3">=1 AND "COL4">=1)
    Statistics
              1  recursive calls
              0  db block gets
           1697  consistent gets
            850  physical reads
              0  redo size
            422  bytes sent via SQL*Net to client
            385  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              1  sorts (memory)
              0  sorts (disk)
              1  rows processed
    SQL> select /*+ index(demo demo_idx)*/ count(distinct col5)
      2   from demo
      3   where col1 between 1 and 50000
      4   and col2=col1
      5   and col3=col1
      6   and col4=col1
      7  /
    COUNT(DISTINCTCOL5)
                      1
    Execution Plan
    Plan hash value: 189561699
    | Id  | Operation                    | Name     | Rows  | Bytes | Cost (%CPU)| T
    ime     |
    |   0 | SELECT STATEMENT             |          |     1 |   116 |   431   (1)| 0
    0:00:06 |
    |   1 |  SORT GROUP BY               |          |     1 |   116 |            |
            |
    |   2 |   TABLE ACCESS BY INDEX ROWID| DEMO     |     1 |   116 |   431   (1)| 0
    0:00:06 |
    |*  3 |    INDEX FULL SCAN           | DEMO_IDX |     1 |       |   430   (1)| 0
    0:00:06 |
    Predicate Information (identified by operation id):
       3 - filter("COL2"="COL1" AND "COL3"="COL1" AND "COL4"="COL1" AND
                  "COL1"<=50000 AND "COL2"<=50000 AND "COL3"<=50000 AND "COL4"<=5000
    0 AND
                  "COL1">=1 AND "COL2">=1 AND "COL3">=1 AND "COL4">=1)
    Statistics
              1  recursive calls
              0  db block gets
          50426  consistent gets
              0  physical reads
              0  redo size
            422  bytes sent via SQL*Net to client
            385  bytes received via SQL*Net from client
              2  SQL*Net roundtrips to/from client
              1  sorts (memory)
              0  sorts (disk)
              1  rows processedSo I don't know what is the default value of dbfmbrc in XE and not gone too deep to understand how for example system statistics may change your situation.
    Gints Plivna
    http://www.gplivna.eu
    P.S. BTW I used Oracle Database 10g Enterprise Edition Release 10.2.0.1.0.
    Message was edited by:
    gintsp
    listened to Williams suggestion :)

  • We have developed a desktop application using Flash software and published the same for MAC environment. When we double click on the index.app file, the application is working perfectly in Mac 10.6.7 at our office. But if the same index.app file is double

    We have developed a desktop application using Flash software and published the same for MAC environment. When we double click on the index.app file, the application is working perfectly in Mac 10.6.7 at our office. But if the same index.app file is double clicked in Mac 10.6.7 at out clients location, its not opening. For your information the client is able to open this same application by double clicking on index.swf file. The main problem is that client is not able to open the application using index.app file at their office whereas we are able to do so at our office. Can anyone give some suggestions to sort this problem?

    The most common reason is different versions of Flash, or different versions of web browsers used.  Some users may elect to not have Flash installed at all because of the processor overhead of Flash.  If you are going to make an application for a client, check what operating system and browser versions they are using first.  Then determine if a stand alone application is required, or if they have the necessary plugins to run specific browser enhanced code.
    P.S. MAC is an acronym for Media Access Control.  Mac is the shorthand for Macintosh, the operating system and computers made by Apple Inc.

  • TS2529 I had backed-up my iPhone and when I try to restore it it says that the back-up is corrupt or the iPhone is not compatible. I had a bunch of pictures that I need from that restore. Is there any way to get them back? I did the back-up to my Macbook

    I had backed-up my iPhone and when I try to restore it it says that the back-up is corrupt or the iPhone is not compatible. I had a bunch of pictures that I need from that restore. Is there any way to get them back? I did the back-up to my Macbook

    If the backup is corrupt, there is nothing that can be done.
    For what reason did you choose to not copy the pictures off the device to a computer prior to attemping an update or whatever caused the need to restore?  iPhones are not intended as sole storage of any media.

  • How do you determine the index of an element in a cluster?

    I've got a small program that I put together, as I'm just trying to get up to speed on Labview.  I have a cluster boolean buttons and for some reason the top buttion is coming through as Index 3.  There are four buttons in the cluster, but for some reason, there's been a shift.  The top one should be Index 0.  How do I get things corrected?  I can't see how to determine the index of each button through Properties.
    Thanks
    Solved!
    Go to Solution.

    You can change the order required by just clicking the element by using a unbundle

  • Query on virtual column that is defined in XMLIndex does not use the index

    Hello,
    I am facing an issue in executing queries on a virtual column that is defined in an XMLIndex: it appears as if the index is not used.
    Database details:
    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    PL/SQL Release 11.2.0.3.0 - Production
    CORE 11.2.0.3.0 Production
    TNS for 64-bit Windows: Version 11.2.0.3.0 - Production
    NLSRTL Version 11.2.0.3.0 - Production
    For this use case the XML documents adhere to the following XSD and are stored in an XMLType column in a table:
    <xsd:schema xmlns:xsd="http://www.w3.org/2001/XMLSchema"
        xmlns="http://a_name_space/v1"
        targetNamespace="http://a_name_space/v1"
        elementFormDefault="qualified" attributeFormDefault="unqualified" version="1.0">
        <xsd:element name="fields">
            <xsd:complexType>
                <xsd:sequence>
                    <xsd:element name="field" maxOccurs="unbounded">
                        <xsd:complexType>
                            <xsd:choice>
                                <xsd:element name="value" minOccurs="1" maxOccurs="1">
                                    <xsd:complexType>
                                        <xsd:simpleContent>
                                            <xsd:extension base="notEmptyString4000Type"/>
                                        </xsd:simpleContent>
                                    </xsd:complexType>
                                </xsd:element>
                                <xsd:element name="values" minOccurs="1" maxOccurs="1">
                                    <xsd:complexType>
                                        <xsd:sequence>
                                            <xsd:element name="value" minOccurs="1" maxOccurs="1">
                                                <xsd:complexType>
                                                    <xsd:simpleContent>
                                                        <xsd:extension base="notEmptyString4000Type">
                                                            <xsd:attribute name="startDate" type="xsd:date" use="required"/>
                                                            <xsd:attribute name="endDate" type="xsd:date" />
                                                        </xsd:extension>
                                                    </xsd:simpleContent>
                                                </xsd:complexType>
                                            </xsd:element>
                                        </xsd:sequence>
                                    </xsd:complexType>
                                </xsd:element>
                            </xsd:choice>
                            <xsd:attribute name="name" type="string30Type" use="required"/>
                            <xsd:attribute name="type" type="dataType" use="required"/>
                        </xsd:complexType>
                    </xsd:element>
                </xsd:sequence>
            </xsd:complexType>
        </xsd:element>
        <xsd:simpleType name="dataType">
            <xsd:annotation>
                <xsd:documentation>Char, Date, Number</xsd:documentation>
            </xsd:annotation>
            <xsd:restriction base="xsd:string">
                <xsd:enumeration value="C"/>
                <xsd:enumeration value="D"/>
                <xsd:enumeration value="N"/>
            </xsd:restriction>
        </xsd:simpleType>
        <xsd:simpleType name="string30Type">
            <xsd:restriction base="xsd:string">
                <xsd:maxLength value="30"/>
            </xsd:restriction>
        </xsd:simpleType>
        <xsd:simpleType name="notEmptyString4000Type">
            <xsd:restriction base="xsd:string">
                <xsd:maxLength value="4000"/>
                <xsd:pattern value=".+"/>
            </xsd:restriction>
        </xsd:simpleType>
    </xsd:schema>A field can have a single value as well as multiple values.
    The XMLIndex is defined as follows:
    CREATE INDEX test_xmltype_idx ON test_xmltype (additional_fields) INDEXTYPE IS XDB.XMLIndex
    PARAMETERS
    XMLTable dt_fld_tab (TABLESPACE "TAB_SPACE" COMPRESS FOR OLTP) ''fields/field''
    COLUMNS
    name varchar2(30 char) PATH ''@name''
    ,dataType varchar2(1 char) PATH ''@type''
    ,val varchar2(4000 char) PATH ''value/text()''
    ,vals XMLType PATH ''values/value'' VIRTUAL
    XMLTable dt_fld_multi_value_tab (TABLESPACE "TAB_SPACE" COMPRESS FOR OLTP) ''value'' passing vals
    COLUMNS
    val varchar2(4000) PATH ''text()''
    ,startDate varchar2(30 char) PATH ''@startDate''
    ,endDate varchar2(30 char) PATH ''@endDate''
    ');The following b-tree indexes are defined:
    create index dt_field_name_idx on dt_fld_tab (name);
    create index dt_field_value_idx on dt_fld_tab (val);
    create index dt_field_values_idx on dt_fld_multi_value_tab (val);And stats are properly computed before the queries are executed:
    call dbms_stats.gather_table_stats(user, 'test_xmltype', estimate_percent => null);Queries for single values are cost efficient and fast. With 600K rows in the table these return with 0.002 seconds.
    Queries for multi-valued fields / elements are not though, these result in a full table scan.
    Sample XML snippet:
    <fields>
      <field name="multiVal" type="C">
        <values>
          <value startDate="2013-01-01" endDate="2013-01-01">100</value>
          <value startDate="2014-01-01">120</value>
        </values>
      </field>
    </fields>Examples of costly and slow queries:
    select id from test_xmltype
    where xmlexists('/fields/field/@name="multiVal"' passing additional_fields)
    and xmlexists('/fields/field/values/value[@startDate="2013-01-01"]' passing additional_fields)
    and xmlexists('/fields/field/values/value[text()="100"]' passing additional_fields)
    select id from test_xmltype
    where xmlexists('/fields/field/@name="multiVal"' passing additional_fields)
    and xmlexists('/fields/field/values/value[@startDate="2013-01-01" and .="100"]' passing additional_fields);Whereas the following query on the multi valued field is fast:
    select id from test_xmltype
    where xmlexists('/fields/field/@name="multiVal"' passing additional_fields)
    and xmlexists('/fields/field/values/value[@startDate="2013-01-01"]' passing additional_fields);For the XPath /fields/field/values/value[@startDate="2013-01-01"] the index is used.
    Suspected cause: XPath issue for the value of a multi valued field, e.g. /fields/field/values/value[text()="aValue"].
    Any hints are appreciated: what am I overlooking here?
    Thanks in advance,
    -Sjoerd
    Edited by: user615230 on May 27, 2013 7:46 AM

    Hello,
    This is using binary XML. The table creation script is:
    create table test_xmltype
    (id number(14,0) not null primary key
    ,member_code varchar2(30 char) not null
    ,period_code varchar2(30 char) not null
    ,amount number(12,2) not null
    ,additional_fields xmltype
    );The schema is not registered in the database. Is that required? It is primarily used to generate Java classes that will be used in order to construct the XML documents.
    And you are right: for our initial investigation the sample XML documents are generated with a PLSQL routine and do not contain namespaces. But for the single valued fields there are also no namespaces and the queries on these are executed with very satisfactory plans.
    Thanks for the swift reply.
    -Sjoerd

  • The Em Key is not configured properly or is corrupted in the file system and doe s not exist in the Management Repository.

    I'm getting this error and is there any way to resolve this without re-creating the service/s again, etc. Just curious if there are any suggestions/steps I may have missed, if not, I'll drop and recreate the services again
    Anyway, I have a backup of the emkey.ora and put it in the sysman/config directory, but still doesn't work... please see below...
    C:\Documents and Settings\abigail>emctl status emkey
    Oracle Enterprise Manager 11g Database Control Release 11.2.0.1.0
    Copyright (c) 1996, 2010 Oracle Corporation.  All rights reserved.
    Please enter repository password:
    The Em Key is not configured properly or is corrupted in the file system and doe
    s not exist in the Management Repository. To correct the problem:
    1) Copy the emkey.ora file from another OMS or backup machine to the OH/sysman/c
    onfig directory.
    2) Configure the emkey.ora file by running "emctl config emkey -emkeyfile <emkey
    .ora file location>".
    C:\Documents and Settings\abigail>emctl config emkey -emkeyfile E:\app\abigail\p
    roduct\11.2.0\dbhome_1\sysman\config
    Oracle Enterprise Manager 11g Database Control Release 11.2.0.1.0
    Copyright (c) 1996, 2010 Oracle Corporation.  All rights reserved.
    Please enter repository password:
    The Em Key is already configured. To override the current configuration, run "em
    ctl config  emkey <options> -force".
    C:\Documents and Settings\abigail>emctl config emkey -emkeyfile E:\app\abigail\p
    roduct\11.2.0\dbhome_1\sysman\config -force -sysman_pwd ********
    Oracle Enterprise Manager 11g Database Control Release 11.2.0.1.0
    Copyright (c) 1996, 2010 Oracle Corporation.  All rights reserved.
    The Em Key could not be configured. Unable to access file E:\app\abigail\product
    \11.2.0\dbhome_1\sysman\config.
    C:\Documents and Settings\abigail>emctl secure dbconsole -sysman_pwd ********
    Oracle Enterprise Manager 11g Database Control Release 11.2.0.1.0
    Copyright (c) 1996, 2010 Oracle Corporation.  All rights reserved.
    http://ABIGAIL:5501/em/console/aboutApplication
    DBCONSOLE already stopped...   Done.
    Agent is already stopped...   Done.
    Securing dbconsole...   Started.
    Please enter repository schema password:
    Securing dbconsole...   Failed.
    EMKey is misconfigured.

    Sorry but I'll have to ignore your suggestion, and don't worry about how I spend my time
    AND OHMYGODD IT WORKSS!!! For future reference, this is how I solved it based on my searches, actually based on one particular search:
    http://franjv.blogspot.com/2013/08/emkey-problems-in-oracle-rac-11gr2.html
    Thank youuu whoever wrote that and... is it the same "Fran" who helped me the last time?!? Coincidence??
    ANyway, thank youuuu

  • Anyone having a problem with the latest number software. Docs do not open Invalid. The index .xml file is missing?

    I have been having problems since updating to the latest version of Numbers.
    When I try to open files I am getting an error.
    "The document “'13 Tax Prep .numbers” is invalid.
    The index.xml file is missing."
    Anyone else getting this message? Anyone know how to fix the issue?

    Jeff Shenk wrote:
    Numbers 3.1 can recognize both the new files without the index.xml file and the old ones with it, which it imports to the new format.
    Agreed on that part.  But take it a step further.  If Numbers 3.1 is asked to open an old document that is still in the old Numbers 2 format that relies on an index.xml file but Numbers 3.1 can't find the index.xml file (perhaps the document is corrupted) then how can Numbers 3.1 open the document and interpret it in order to convert it to the new format? My guess is that this a possible trigger for the 'index.xml missing' error message.
    Interesting link here on the new document format.
    SG

Maybe you are looking for

  • Internal Hard Drives for Apple Mac Pro

    Hello there, Does any one have some advice on the following internal HDs. I am in Germany and this is what I have found that I am interested in. I just don't know if they all are for Apple Mac Pro because the provider doesn't specify. 1) HD 1000GB We

  • SCORE: Polyphonic Score Style & Rests

    Hello Score Gurus, This horn part is written within a single MIDI region using a polyphonic score style with four parts (on 4 MIDI channels) to allocate the notes to the proper voice between the two staves. In this example, only horns 1 and 3 are pla

  • Embedded LDAP

    What's inside server_inst/ldap/ldapfiles/EmbeddedLDAP? Is there anyway to refresh this file so it can clean all the old data? This is the problem: A role is not currently assigned to a user. But in the bean code, when this client with the pricipal(us

  • How do I delete this unwanted U2 download that Apple have stuck on my ipad

    HHow to delete unwanted downloads such as the rubbish U2 album

  • Import folder as if it's a CD

    Hi, I have folders of music (archived from LP) that I want to import. when I import a CD I can have the Title, year, artist, etc., information for all those tunes entered in the Get Info for the disc and then have them show up for all the tunes. Is t