Mutating table and row state - this was unexpected

So, I learned in a class about 3 years ago to expect the following
SQL> create table p (pk number primary key);
Table created.
SQL> create table c (fk number references p(pk));
Table created.
SQL> create or replace trigger t_insrt
  2  before insert on p
  3  for each row
  4  begin
  5   insert into c values (:new.pk);
  6  end;
  7  /
Trigger created.
SQL> insert into p values (1);
insert into p values (1)
ERROR at line 1:
ORA-02291: integrity constraint (FORBESC.SYS_C00169150) violated - parent key
not found
ORA-06512: at "FORBESC.T_INSRT", line 2
ORA-04088: error during execution of trigger 'FORBESC.T_INSRT'and so it led me to think that replicating ON MODIFY PARENT - MODIFY CHILD functionality wouldn't work in a BEFORE ROW trigger, but it does
SQL> drop trigger t_insrt;
Trigger dropped.
SQL> create or replace trigger p_updt
  2  before update on p
  3  for each row
  4  begin
  5   update c
  6   set fk = :new.pk
  7   where fk = :old.pk;
  8  end;
  9  /
Trigger created.
SQL> insert into p values (1);
1 row created.
SQL> insert into c values (1);
1 row created.
SQL> select * from c;
        FK
         1
SQL> update p
  2  set pk = 2
  3  where pk = 1;
1 row updated.
SQL> select * from c;
        FK
         2Why would the first scenario fail while the second succeeds? The update seems prone to a parent record also not existing, at least not by the BEFORE ROW trigger.
---=Chuck

< mutating table and row state >
BTW, you don't seem to have run into the mutating table error though 2 other threads today are also about it. You have a constraint violation, a different thing entirely.
I believe the second scenario works because you're neatly avoiding the error of the first. The error "ORA-02291: integrity constraint (FORBESC.SYS_C00169150) violated" means that on insert Oracle is looking up the value you're trying to insert, not finding it, and raising an error. With the before trigger you are taking the assigned value from the insert, updating the parent to it, so that on actual insert when the check happens the value is there due to the update.
I'm not convinced this is a good idea because any on-the-fly approach to data entry needs to be examined carefully.

Similar Messages

  • Every time I try to sign on it says that ichat lost the connection to the AIM account and the server connection was unexpectedly lost.  How can I fix this?

    Every time I try to sign on it says that ichat lost the connection to the AIM account and the server connection was unexpectedly lost.  How can I fix this?

    Note For others.
    For the Port Changes (or server names in some cases) you have to be logged out to make changes on the Server Settings pane.
    9:41 PM      Friday; October 28, 2011
    Please, if posting Logs, do not post any Log info after the line "Binary Images for iChat"
     G4/1GhzDual MDD (Leopard 10.5.8)
     MacBookPro 2Gb( 10.6.8)
     Mac OS X (10.6.8),
    "Limit the Logs to the Bits above Binary Images."  No, Seriously

  • Mutating table and recursive triggers

    Hello every one,
    I seem to be having a problem with a statement level update trigger. I want to use a trigger on a table then update the same table. Initially when I coded it as a row trigger but I received a ORA 04091 mutating error message.
    I refered a Oracle PL/SQL book and it suggested a work around. I recoded the trigger but it is not working. Oracle does not want to let me update the table I am running my statement level trigger(accompanied with a package and row level trigger) on. Please see code below. CREATE OR REPLACE PACKAGE service_guarantee_cap_data AS
    TYPE t_userid is table of bizguaranteereq.userid%type
    index by BINARY_INTEGER;
    TYPE t_recordid is table of bizguaranteereq.recordid%type
    index by BINARY_INTEGER;
    v_userid t_userid;
    v_recordid t_recordid;
    v_NumEntries BINARY_INTEGER :=0;
    end service_guarantee_cap_data;
    CREATE OR REPLACE TRIGGER service_guar_cap_row_update
    BEFORE UPDATE of statuscd ON bizguaranteereq
    FOR EACH ROW
    BEGIN
    service_guarantee_cap_data.v_NumEntries := service_guarantee_cap_data.v_NumEntries + 1;
    service_guarantee_cap_data.v_userid(service_guarantee_cap_data.v_NumEntries ):= :new.userid;
    service_guarantee_cap_data.v_recordid(service_guarantee_cap_data.v_NumEntries ) := :new.recordid;
    END service_guar_cap_row_update;
    CREATE OR REPLACE TRIGGER service_guar_cap_stat_update
    after UPDATE of statuscd ON bizguaranteereq
    DECLARE
    v_rowid_2 number;
    v_userid_2 number;
    v_recordid_2 number;
    v_overcap_99 number;
    BEGIN
    v_0 := 0;
    v_overcap_99 := 99;
    FOR v_loopindex in 1..service_guarantee_cap_data.v_NumEntries LOOP
    v_userid_2 := service_guarantee_cap_data.v_userid(v_loopindex);
    v_recordid_2 := service_guarantee_cap_data.v_recordid(v_loopindex);
    select count(*)
    into v_statuscd_count
    from bizguaranteereq
    where userid=v_userid_2
    and statuscd=99;
    IF v_statuscd_count > 0
    THEN
    update bizguaranteereq
    set statuscd=v_overcap_99
    where userid = v_userid_2
    and recordid=v_recordid_2;
    END IF;
    END LOOP;
    service_guarantee_cap_data.v_NumEntries := v_0;
    END service_guar_cap_stat_update;

    I did not receive a mutating error with my work around code. I think the problem lies with the update DML. I will try your suggestion of an external proceedure that takes the parameter given from the statment level trigger and then updates the table using the proceedure.
    Thanks
    <BLOCKQUOTE><font size="1" face="Verdana, Arial, Helvetica">quote:</font><HR>Originally posted by wasim hamwi ([email protected]):
    I am not sure this code will work, you will get mutating table because you are selecting the count(*) from the sam table you are tying to update....use Autonomous_transaction if your Db iv V8.1.5 or higher.
    Wasim<HR></BLOCKQUOTE>
    null

  • How to partition tables and indexes in this scenario?

    So our situation is pretty simple. We have 3 tables.
    A, B and C
    the model is A->>B->>C
    Currently A, B and C are range partitioned on a key created_date however it's typical that only C is every qualfied with created date. There is a foreign key from B -> A and C -> B
    we have many queries where the data is identified by state that is indexed currently non partitioned on columns in A ... there are also indexes on the foreign keys that get from C -> B -> A. Again these are non partitioned indexes at this time.
    It is typical that we qualify A on either account or user or both. There are indexes (non partitioned) on these
    We have a problem now because many of the queries use leading wildcards ie. account like '%ACCOUNT' etc. This often results in large full table scans. Our solution has been to remove the leading wildcard but this isn't always possible.
    We are wondering how we can benefit from partitioning and or sub partitioning table A. since it's partitioned on created_date but rarely qualify by that.
    We are also wondering where and how we can benefit from either global partitioned index or local partitioned indexes on tables A. We suspect that the index on the foreign key from C to B could be a local partitioned index.
    I am also wondering what impact pushing the state from A that's used to qualify A down to C would have any advantage.
    C is the table that currently we qualify with the partition key so I figure if you also pushed down the state from A that's used to qualify the set of C's we want based on the set of B's we want based on the set of A thru qualfying on columns within A.
    If we push down some of those values to C and simply use C when filtering I'm wondering what the plans will look like compared to having to work all the way up from the bottom to the top before it begins qualifying A.
    Edited by: steffi on Jan 14, 2011 11:36 PM

    We are wondering how we can benefit from partitioning and or sub partitioning table A. since it's partitioned on >created_date but rarely qualify by that. Very good question. Why did you partition on it? You will never have INSERTS on these partitions, but maybe deletes and updates? The only advantage (I can think of) would be to set these partitions in a read only tablespace to ease backup... but that's a weired reason.
    we have many queries where the data is identified by state that is indexed currently non partitioned on columns in >A ... there are also indexes on the foreign keys that get from C -> B -> A. Again these are non partitioned indexes at >this time.Of course. Why should they be partitioned by Create_date?
    It is typical that we qualify A on either account or user or both. There are indexes (non partitioned) on these
    We have a problem now because many of the queries use leading wildcards ie. account like '%ACCOUNT' etc. This >often results in large full table scans. Our solution has been to remove the leading wildcard but this isn't always possible.I would suspect full index scan. Isn't it?
    We are also wondering where and how we can benefit from either global partitioned index or local partitioned >indexes on tables A. We suspect that the index on the foreign key from C to B could be a local partitioned index.As A is not accessed by any partition, why should C and B profit? You should look to partition by the key you are using to access. But, you are looking to tune your SQLs where the access is like '%ACCOUNT' on A. Then when there is a match. ORACLE joins via your index and nested loop (right?) to B and C.
    I am also wondering what impact pushing the state from A that's used to qualify A down to C would have any >advantage.Why should it. It just makes the table and indexes larger => more IO.
    C is the table that currently we qualify with the partition key so I figure if you also pushed down the state from A >that's used to qualify the set of C's we want based on the set of B's we want based on the set of A thru qualfying >on columns within A.If the access from A to C would be .. AND A.CREATE_DATE =C.CREATE_DATE and c.key like '%what I want%' which does not qualifify for a FK ;-) then, as that could be resulting in a partition scan, you could "profit". But, I'm sure that's not your model.
    If we push down some of those values to C and simply use C when filtering I'm wondering what the plans will look >like compared to having to work all the way up from the bottom to the top before it begins qualifying A.So you want to denormalize A,B,C and into one table? With the same access is like '%ACCOUNT' you would get a full scan on an index as before, just the objects would be larger due to redundance and harder to maintain. In the end you would have a bad and slower design.
    Maybe you explain what the problem is.
    Full index scan can not be avoided, but that can be made faster by e.g. parallel query, and then the join to B and C should be a "snip" if you just identify a small subset of rows in these tables.

  • JSP: DispalyTag - Nested table and rows

    Hi!
    I�d like to know if it�s possibile create new row with dispalytag.
    I have a nested table (and long list), but I�d like to have the items under the header, not items in a column of the header. I tried with <tr> but doesn�t work.
    Is it possibile to creat a row? And how??? I thought I must use decorators (org.displaytag.model.Row) but when I run my application it doesn't work.
    Can you help me?
    Thanks
    Annalisa

    Thanks for the quick response. I was thinking I might get a response from someone who had tried it, so I didn't waste time trying it myself. I consider your response is pretty authoritative, and I'll take that as a sign that I should forget the nested tables design.
    No, polymorphic views probably won't do the job. I think I'll just go ahead and create staging versions of each detail table with foreign keys back to the master staging table. Then I'll let the wizard create the needed association objects and view links in ADF BC. It will complicate the procedure for applying changes, particularly adding a row to the change history table. But that's the price we'll need to pay.
    OTOH - I just had a thought - since the change history table is mostly for auditing, it appears in some reports, but has no CRUD in the application, other than the insert when the changes are applied. If I did that insert with a trigger or some other PL/SQL, then the change history could still be a single table with nested table columns for the details.

  • If you have an itunes acct that was not set up to share and the lap top it was on is no longer working, is there a way to open up the old itunes acct to get music and photos?  This was before icloud was available, can you reset your password?

    If you have an itunes acct that was not set up to share and the lap top it was on is no longer working, is there a way to open up the old itunes acct to get music and photos off of it?  This was before Icloud was available.  Can you reset your itunes password somehow?

    iTunes and the iTunes Store won't have anything to do with your photos. Those would be stored only on your laptop unless you made a backup of some sort. If you didn't, and the hard drive in your laptop has failed, then your photos are gone unless you wish to pay a disk recovery service a lot of money to try and get them back.
    As to your music, if you purchased it from the iTunes Store you can probably re-download it. Go to the iTunes Store, log into your account, and click the Purchases link under the Quick Links. From there you should be able to re-download some or all of your purchased content. Note that not all content has been licensed for re-downloading in all countries at this time. You can see what content you can download here:
    http://support.apple.com/kb/HT5085
    You can also re-download content using an iOS device.
    For full instructions, see:
    http://support.apple.com/kb/ht2519
    Regards.

  • Why is iphoto unexpectedly quitting when trying to order prints?  I called yesterday and apple indicated this was a problem.  Any idea when the problem will be fixed?

    I photo is unexpectedly quitting when placing an order.

    That's what happens if you read 12 month old threads
    http://support.apple.com/kb/HT1564
    Regards
    TD

  • Self referencing table and SQL statement

    In my database, I have a self-referencing table, the table itself is for projects, and it allows users to get a hierarchical view of the company.
    Here is the SQL (modifier is the term we use for project code, BBCI is the top project)
    SELECT
    modifier, modifierDescription, level
    FROM
    modifier
    WHERE
    level <= 2
    CONNECT BY PRIOR
    modifier = parentModifier
    START WITH modifier = 'BBCI'
    ORDER BY level
    That perticular query gets the first two levels in the structure. I use this information to produce a tree structure in a web app.
    But users have requested it would be good if in the tree structure is showed an + or - depending on whether there were anymore children under each parent, or better still the number of children under it, for example
    BBCI
    + BBCI_CHILD
    + BBCI_CHILD2
    - BBCI_CHILD3
    or
    BBCI
    + BBCI_CHILD (3 projects underneath)
    + BBCI_CHILD2 (2 projects underneath)
    - BBCI_CHILD3 (0 projects underneath)
    I am really stumped on this issue, and I am sure there is a way to do this in the web app, so for example do a query for each child node to see how many child nodes are underneath, but I figure it would be a lot tidier and faster if I could do it from a single SQL statement. Unfortunately I have tried to do this and am very much stuck.
    If you need more information please let me know
    Thanks!
    Jon

    You may be able to do this using analytical functions but it depends on the Oracle version you are using. It can also be done with standard SQL - you just need to count the number of child rows for each modifier/project first and supply that list as an in-line view:
    SELECT decode(modifier,'X',null,decode(child_rows,null,'-','+')),
    m.modifier,
    decode(modifier,'X',null,'('||nvl(cq.child_rows,0)||' projects underneath)')
    FROM modifier m,
    (select parentModifier,
         count(parentModifier) child_rows
    from modifier
    where parentModifier is not null
    group by parentModifier) cq
    WHERE m.modifier=cq.parentModifier(+)
    AND level <= 2
    CONNECT BY PRIOR
         m.modifier = m.parentModifier
    START WITH modifier = 'X'
    ORDER BY level, modifier;
    which gives you something like...
    D MODIFIER DECODE(MODIFIER,'X',NULL
    X
    - Y1 (0 projects underneath)
    + Y2 (2 projects underneath)
    + Y3 (3 projects underneath)
    The decode stuff is just to show you the result, you can format the output in your calling code. Just extend the columns to include everything you want, modifierDescription etc..
    Hope this helps.

  • Finding Table and Row size in bytes

    Hi Guys,
    I want to find how much space a TABLE is occupying in bytes and also how much space 1 individual ROW is occupying in a data block size of 8192 bytes.
    Thank you.
    Regards,
    Kamran

    Hi for calculating size of the table you can use
    select sum(bytes) from dba_segments
    where segment_name = 'YOUR_TABLE'and for rowsize you can use something like this
    select 
         max(vsize(empno)) +
         max(vsize(ename)) +
         max(vsize(job)) +
         max(vsize(mgr)) +
         max(vsize(hiredate)) +
         max(vsize(sal)) +
         max(vsize(comm)) +
         max(vsize(deptno))  Bytes
    from
         scott.emp;

  • Number of records are different in database table and select statement

    Hi All,
    i need to fetch data from table BSID for the customer 0010000145
    if i am writing the code like -
        SELECT bukrs kunnr umskz shkzg dmbtr zfbdt zbd1t kkber
               FROM bsid
               INTO TABLE it_bsid
               FOR ALL ENTRIES IN it_kna1
               WHERE kunnr = it_kna1-kunnr
               and bukrs = pa_bukrs.
    no. of records are 130 in the internal table it_bsid
    and actual records are 200
    but when i am hardcoding the customer no.
    i am getting the exact records
       SELECT bukrs kunnr umskz shkzg dmbtr zfbdt zbd1t kkber
              FROM bsid
              INTO TABLE it_bsid
              FOR ALL ENTRIES IN it_kna1
              WHERE bukrs = pa_bukrs
                AND kunnr = '0010000145'.
    records in internal table = records in the database = 200.
    how it is possible.
    why the first code is not giving the correct no. of records.
    please reply asap.
    thanks in advance,
    madhu

    Madhu,
    You need to use the conversion routine...before u pass KUNNUR.
    CONVERSION_EXIT_ALPHA_RANGE_I
    CONVERSION_EXIT_ALPHA_RANGE_O
    Hope this helps..
    Chandra.

  • New to Apple and I thought this was supposed to be easy!!!

    I finally left the darkside and joined the force but im mad as H#@&! My macbook airport and internet connection say im connected to my linksys wireless router and even shown connection to my network but i still cant connect to the net via safari. I am currently using WPA and have accurately input my security code. I also disabled the security but nothing changed. if i wanted this, i could have stayed with windows...

    Hello dj mark h. Welcome to the Apple Discussions!
    Let's double-check your AirPort settings on the MacBook ...
    System Preferences > Network > AirPort
    o Location: Automatic
    o In the left column, verify that an AirPort option exists. If it does not, click "+" to create one. (Note: If you are unable to create an AirPort configuration, the AirPort card in your computer either doesn't exist or it is not recognized.)
    o Verify that AirPort is "On." Click "Turn AirPort On" button if not.
    o Ask to join new networks: (enabled)
    o Show AirPort status in menu bar: (enabled)
    Systems Preferences > Network > AirPort > Advanced
    AirPort tab
    o Remember any network this computer has joined: (enabled)
    o Disconnect from wireless networks when logging out: (enabled)
    o Require Administrator password to control AirPort: (disabled)
    TCP/IP tab
    o Configure IPv4: Using DHCP
    o Configure IPv6: Automatically
    DNS tab
    o DNS Servers: <if required, add your desired DNS server IP address(es) here>
    o Search Domains: <if required, add your desired DNS Search Domain(s) here>
    WINS tab
    o Enter desired WINS info or leave the defaults
    AppleTalk tab
    o Make AppleTalk Active: (disabled)
    802.11X tab
    o Enter desired 802.11X info or leave the defaults
    Proxies tab
    o Enter the desired Proxies info or leave the defaults

  • On my itunes (on pc) and on my touch i had many songs and albums, purchased, burnt and transfered.  this was fine and always synchronized both ways. had problems with my hard drive and had to reformat it all. since then my pc itunes only got back some 60%

    when i try to synchronize, it says that it would first erase all in the ipod before syncing.  all the original contents are intact on the ipod touch.  i have backed it up to the pc, yet, to sync. it insists on first erasing all.     please someone help

    Follow these instructions
    Syncing to a "New" Computer or replacing a "crashed" Hard Drive: Apple Support Communities

  • How to fix skewed table columns and rows after re-import XML

    My question is regarding XML Import in InDesign CS3.
    I have a XML that has a table of 5 columns and 5 rows, when I import it into InDesign, the table shows up fine with 5 columns and 5 rows. However when I revise my table to have an additional column, and re-import the XML file, the table gets updated, but instead of with an additional column, it gets 'appended' with an additional row instead (InDesign seems to blindly replace each cell from left to right, top to bottom, and ends up with 6 rows instead). The XML file specifies the number of columns and rows (5 and 5 before, 5 and 6 after), why doesn't InDesign recognize it and automatically add a new column when I re-import the file?  Is this problem fixed in CS5.5? Is there a script to fix this?
    Here is an example of my old XML vs new XML:
    Old:
    <frame5>
    <Table xmlns:aid="http://ns.adobe.com/AdobeInDesign/4.0/" aid:table="table" aid:trows="5" aid:tcols="5">
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1"></Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2006</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2005</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2004</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2003</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Stores at beginning of period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">846</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">805</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">753</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">678</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Stores opened during the period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">36</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">50</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">59</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">79</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Stores closed during the period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">13</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">9</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">7</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">4</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Total stores at end of period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">869</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">846</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">805</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">753</Cell>
    </Table>
    </frame5>
    New:
    <frame5>
    <Table xmlns:aid="http://ns.adobe.com/AdobeInDesign/4.0/"
    xmlns:aid5="http://ns.adobe.com/AdobeInDesign/5.0/" aid:table="table" aid:trows="5" aid:tcols="6">
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1"></Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2007</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2006</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2005</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2004</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">2003</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Stores at beginning of period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">123</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">846</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">805</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">753</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">678</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Stores opened during the period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">456</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">36</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">50</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">59</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">79</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Stores closed during the period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">789</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">13</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">9</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">7</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">4</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">Total stores at end of period</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">1368</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">869</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">846</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">805</Cell>
    <Cell aid:table="cell" aid:crows="1" aid:ccols="1">753</Cell>
    </Table>
    </frame5>

    What I mean is, right now there is a "workaround" which requires a user to manually add an extra column before re-importing the XML (with an extra column), otherwise the results get skewed. If by providing a script we can simply ask the user to "run" it, it would be a more "acceptable" solution.  
    Right. So, one solution would be to use Convert Text to Tables. If you find that works for you, then you can script it.
    That's why I asked you...
    Of course if nothing is required from the user other than simply re-importing, then that would be the best solution.
    Well, one could imagine a script that was attached to the re-import command, or to the link update notification, but probably the first step is to get something that works reasonably well without automating it completely. Especially because triggers to run scripts silently can introduce hard-to-debug problems.
    Sure we can switch to use "CALS table and see if this works. The question is, why should we need to? In the 2nd XML there is clearly a different "aid:tcols" value, and yet InDesign seems to ignore it and assume the same # of columns? This sounds like an Indesign bug, can someone confirm? Is there any plans to fix this?
    Not to be snarky, but do you want it to work or not? There are the tools you have and the tools you wish you had, and you can't really do much with the ones you wish you had. I'm kind of assuming you are looking for a solution today, not a solution in 2013.
    Yes,  I believe two of us have confirmed this appears to be a bug.
    Plans to fix? Well, we can't really tell you. You could try asking Adobe, but that's not easy information to get out. But you can certainly open a support case at http://adobe.com/go/supportportal and ask. It's not like we have special information here...
    But you're probably better off filing a bug first, in that same fashion.
    But let's assume no one had filed the bug. CS6 is expected in the March/April 2012 timeframe. That means that they're probably just putting the finishing touches on it right now, and it's going to be very difficult to fix things in it now. So then the earliest you'd likely get it fixed in CS6.5/CS7/whatever which presumably comes out by 2Q2013, and that's assuming Adobe decides to fix it...
    I also can't find much documentation on how to update my table to a "CALS table", any examples?
    Try this thread:
    Re: Importing a CALS table into InDesign CS3

  • Extract is picking up rows twice because both an explicit TABLE and a TABLE

    Our extract parameter file has this following(full file uploaded):
    TABLE TK_APP_DATA.PAYMENT_INFO , COLSEXCEPT (ENCRYPTED_VALUE,HASH_VALUE)
    #add_scn()
    TABLE TK_APP_DATA.*
    #add_scn()
    When i look at the change data inserted by the replicat on target and at the extract trail file I can see that a row which is updated is actually extracted twice and duplicated. I also see that the table is resolving twice in the EXTRACT rpt file:
    Using the following key columns for source table TK_APP_DATA.PAYMENT_INFO : INSTRUMENT_ID, FIELD_HANDLE.
    TABLEWildcard resolved (entry SCECOMM.*):
    TABLE SCECOMM.PAYMENT_INSTRUMENT_FIELD, TOKENS ( TK-SCN = @GETENV ( "TRANSACTION" , "CSN" ) ) ;
    Using the following key columns for source table TK_APP_DATA.PAYMENT_INFO : INSTRUMENT_ID, FIELD_HANDLE.
    TABLEWildcard resolved (entry SCECOMM.*):
    TABLE SCECOMM.APPLICATION_PARAMETER, TOKENS ( TK-SCN = @GETENV ( "TRANSACTION" , "CSN" ) ) ;
    Anyone know how I can make it so that only one TABLE statement is used to resolve it? Or conversely a way to filter out table in a wild card statement? I would prefer to avoid having to create a seperate extract for just 1 table(schema has over 350+ tables in it)
    I tried using a TABLEEXCLUDE but that excludes the table entirely. I also tried using WILDCARDRESOLVE IMMEDIATE on the specific table mappings, but can't because we also use DDL replication.
    Any help would be greatly appreciated.
    #anh

    This is unfortunately what I have right now as my stop gap. We will sql generate the list of tables with every deployment which with our agile development is like every other week haha. I do have a requirement to replicate dynamically(the source side changes extremely frequently by adding and dropping tables) but even after talking to oracle support this is not possible. If you have a custom mapping for a table and use a wild card later which would resolve that table you will always get duplicate rows extracted.
    I talked to oracle support(with raised this as a bug/feature request) and raised this with our oracle engagement manager and asked them to add better functionality when it comes to filtering and wild card selection for TABLE and MAP statements to have an option setting to only map once. It's in their queue however not sure how high of a priority it will be for them(I'm guessing not very high =) )

  • Why there is a huge difference between a row size on a disk based table and a row size in memoptimized table of SQL 2014?

    Hi All,
    I have two table with similar structure and data, one is on disk and the other is in memory. I somehow calculated the difference between a row size of on disk and in memory table and found that the row size of in memory is 700 Bytes more than the disk based
    tables.
    aa

    As others mentioned, memory optimized tables and disk based tables have different structures in SQL Server 2014.
    For memory optimized tables, the number of indexes on table also contribute to the size. You can calculate the exact size of rows and thus the table size using the formula given in the below articles
    Table and Row Size in Memory-Optimized Tables
    Estimate the Size of a Table
    Krishnakumar S

Maybe you are looking for

  • Mixing RAM Speeds

    Hi there, I posted this in the "Using Mac" forum, but unsure how to delete it, so I reposted here! I'd like to upgrade my memory. Currently, I have the factory installed 1GB RAM. Can I mix memory speeds? For example, will I be able to put 2GB in the

  • XML DB when you have a lot of XSDs....

    My client has a data situation where we have a lot of XSD schemas (and multiple versions of each XSD) which describe data that needs to be loaded. If I wanted to use schema-based / structured-storage....Can the same xmltype column be mapped to multip

  • NWDI connection with NWDS fails with CIM_ERR_FAILED

    Hello, I tried to import a development configuration from our NWDI. Before that I configured the URL in the preferences and could successfully ping the server. When I try to import a development configuration in the Development Configurations Perspec

  • Is there a way to rebuild the database(s) in ARD without re-installing?

    On my intel iMac (10.4.7) my lists will all go offline or "--" all at once... a restart of the app sometimes solves this for a while but eventually, I need to completely delete all of ARD and re-install to restore functionality. Id there a way to reb

  • Consistent crash when adding PIP

    iMovie keeps crashing when trying to add a photo in the movie as PIP. Any idea? Well, I was going to add the crash report, but I lost it...