Rows per batch and Maximum insert commit size

Have been reading various post to understand what the properties 'Maximum insert commit size' and 'Rows per batch' actually do and am confused as each have their own explanation to these properties. As far as I understood, MICS helps SQL server engine to
commit the number of rows specified to it and RPB is to define the number of rows in a batch?
In case I set the property to say RPB - 10000 and MICS - 5000, does this mean the input data is fetched in a batch of 10000 and each time only 5000 are committed?
What happens in case of RPB - 5000, MICS - 10000. Are the batches formed for each 5000 records and committed for each 10k records?
One post mentioned on RPB being merely a property to help SQL server device a optimal plan for data load. If this is true then why have the MICS property? Instead RPB can be assigned a value and engine decides on what is best load method it wants to use.
Thanks!
Rohit 

Hi Rohit,
Maximum insert commit size specify the batch size that the component tries to commit during fast load operations. The value of 0 indicates that all data is committed in a single batch after all rows have been processed. A value of 0 might cause the running
package to stop responding if the component and another data flow component are updating the same source table. To prevent the package from stopping, set the Maximum insert commit size option to 2147483647.
Rows per batch specify the number of rows in a batch. The default value of this property is –1, which indicates that no value has been assigned.
For more details about those two properties, please refer to the following document:
http://msdn.microsoft.com/en-IN/library/ms188439.aspx
Thanks,
Katherine Xiong
Katherine Xiong
TechNet Community Support

Similar Messages

  • Rows per batch and DefaultbufferMaxrows

    Hi,
    What is the difference between rows per batch and DefaultbufferMaxrows in data task? Is there any relation between these two?

    Rows per batch : It is the number of rows fetched under single operation to optimize the query plan.The rows per batch should be as close to to total amount of records in your source table.
    The default value for this setting is -1 which specifies all incoming rows will be treated as a single batch. You can change this default behavior and break all incoming rows into multiple batches. The allowed value is only positive integer which specifies
    the maximum number of rows in a batch.
    DefaultBufferMaxRows : Defines the maximum number of rows in each buffer, by setting the
    DefaultBufferMaxRows property. The default maximum number of rows is 10,000.
    References:
    http://www.mssqltips.com/sqlservertip/1840/sql-server-integration-services-ssis-best-practices/
    http://social.msdn.microsoft.com/forums/sqlserver/en-US/86efc54e-45bd-4ccc-baf1-a7b34f80fe32/rows-per-batch-property
    If this post answers your query, please click "Mark As Answer" or "Vote as Helpful".

  • Set Maximum Rows Per Page not working

    I want to display 50 rows per page and I tried the below but still I can see the default 15 rows per page.
    Home>Application Builder>Application 301>Page 51>Interactive Report Attributes>Search Bar>Maximum Rows Per Page
    Set to = 50
    Home>Application Builder>Application 301>Page 51>Interactive Report Attributes>Search Bar>Rows Per Page Selector
    Removed the check mark and still shows 15 rows per page.

    Run your report. Set the Rows to 50. Press Go. Select 'Save Report' from the 'magic wheel' menu and next select 'Save as default report settings'.
    It is a rather long way, but seems the only one...
    Roel

  • How to pivot horizontally Author names, and group by Book title. One row per book with multiple authors

    I have 3 tables - Book, Author, BookAuthorReference
    A book can have multiple authors, and when I do straight query I get multiple rows per book
    SELECT <columns>
    FROM Book b, Author a, BookAuthorReference ba
    where ba.BookId = b.BookId and
    ba.AuthorId = a.AuthorId
    I want to get the results as ONE row per book, and Authors separated by commas, like:
    SQL 2008 internals book    Paul Randal, Kimberly Tripp, Jonathan K, Joe Sack...something like this
    Thank you in advance

    This can by done by straying into XML land. The syntax is anything but intuitive, but it works. And moreover, it is guaranteed to work.
    SELECT b.Title, substring(a.Authors, 1, len(a.Authors) - 1) AS Authors
    FROM   Books b
    CROSS  APPLY (SELECT a.Author + ','
                  FROM   BookAuthorReference ba
                  JOIN   Authors a ON a.AuthorID = ba.AuthorID
                  WHERE  ba.BookID = a.BookID
                  ORDER  BY ba.AuthorNo
                  FOR XML PATH('')) AS a(Authors)
    Erland Sommarskog, SQL Server MVP, [email protected]

  • Limit rows per page, include spacer and format last row on page break

    Hello,
    I am new to BI and would like some help with an issue I don't seem to be able to resolve?
    I have a table that uses two xml group elements (2nd nested inside first) to produce my detail lines.
    Example:-
    First for-each loop runs over shipto details and within this another for-each loop runs over items related to that shipto, all of which results in multiple detail lines.
    I need to limit the number of rows per page and should the data roll over to the next page I need to place a horizontal closing border on the first page's table (only using column borders).
    Also when the data does not fill the page I need to add spacer rows to the page limit?
    I have tried using examples found in these forums, position(), counters etc… but due to having the second detail loop I’m unable to get these working or I’m doing something wrong?
    Your help would be greatly appreciated.

    Hi BIPuser, thanks for the links.
    I have tried some of the examples which I can get working with a single level of detail lines. However when I add the second level of detail lines the position() function seems to fail as my rowlimit appears to be ignored.
    I am also unable to get the $lpp example working because the outer detail loop may only have one xml element and therefore BI assumes only one line is to be printed when in fact the inner xml element detail loop may have further multiple lines. I need to effectively count the outer & inner loops as one to determine the number of lines being processed.
    Example of xml - My table for detail lines needs to include both ShipTo & item number lines grouped as shown, thereby counting & outputting 7 lines in this instance.
    <INVOICE>
    <On_Ship_To_S23> Outer loop
    <On_itm_num_shr_row_S21> Inner Loop
    <On_itm_num_shr_row_S21>
    <On_itm_num_shr_row_S21>
    </On_Ship_To_S23>
    <On_Ship_To_S23>
    <On_itm_num_shr_row_S21>
    <On_itm_num_shr_row_S21>
    </On_Ship_To_S23>
    </INVOICE>

  • Delete and Re-Insert a Row in the Same Statement

    Hello to all,
    I have to do a row Update of 2-3 columns, but due to my Triggers(delete and Insert) and unique index, I would like for same row one delete and one Insert of the columns modified,
    is it possibile in same(single) statement do this?
    That is,instead to do
    Update table1 set col3 = '6';
    I should do
          DELETE FROM  table1
          WHERE
    INSERT INTO table1
          SELECT col1,col2  ,'6'
         FROM table1
          WHERE -----thanks in advance
    Edited by: Rosario Vigilante on Mar 5, 2012 11:18 AM

    first of all thanks for your answer and your time,
    I already have 2 Triggers (OF delete and insert) that work well, when they fire they run a stored procedure that inserts rows into another table,
    ie, both triggers on the table 1 write rows on Table 2, more or less what happens in an audit table.
    Anyway thanks again

  • OCI - Array-Fetch vs. One-row-per-fetch

    Hello guys,
    i have a question about the OCI and the possibilities about fetches.
    Is it possible to get only one row per fetch without setting the array size to 1?
    I have a third party application that shows this behaviour which i have rebuild in sqlplus.
    SQL> create table mytest (a number);
    SQL> begin
      2  for i in 1 .. 1500 loop
      3  insert into mytest values (i);
      4      end loop;
      5  end;
      6  /
    PL/SQL procedure successfully completed.
    SQL> commit;
    Commit complete.
    SQL> set autotrace traceonly;
    -- Now with the default array size of 15 with sqlplus
    SQL> select * from mytest;
         118  consistent gets
         101  SQL*Net roundtrips to/from client
         1500  rows processed
    -- Now with a bigger array size (150)
    SQL> set arraysize 150
    SQL> select * from mytest;
         17  consistent gets
         11  SQL*Net roundtrips to/from client
         1500  rows processed
    -- Now the behaviour of the third party application
    SQL> set arraysize 1
    SQL> select * from mytest;
         757  consistent gets
         751  SQL*Net roundtrips to/from client
         1500  rows processed
    SQL> set arraysize 2
    SQL> select * from mytest;
         757  consistent gets
         751  SQL*Net roundtrips to/from client
         1500  rows processedThe third party application is a c program and i can not take a look at the code.
    So as you can see the consistent gets are the same with arraysize 1 and 2. The sql statement which is executed of the c-program is returning a huge amount of data and it seems like it is run with arraysize 1 or 2 or it is executing a different OCI call.
    So now is my question:
    Which methods does the OCI interface provide to recieve (fetch) data?
    - Is it only array fetching (like sqlplus do) or is it possible to return only one row per fetch with a specific call.
    I can speed up the query by setting the bigger array-size in sqlplus .. but i want to point the programers to the possibilities with the OCI.
    Thanks and Regards
    Stefan

    The following call in OCI can be used to control the fetched rows
    MAX_PREFETCH_ROWS is number of rows you want to fetch in one round trip.
    (void) OCIAttrSet((dvoid *)DBctx->stmthp, (ub4) OCI_HTYPE_STMT,
    (dvoid *)&MAX_PREFETCH_ROWS,(ub4)sizeof(MAX_PREFETCH_ROWS),(ub4) OCI_ATTR_PREFETCH_ROWS, DBctx->errhp);

  • FTP adapter Batching is not inserting records in to  multiple tables

    Hi,
    Env : 11.1.1.5
    We have a scenario where we are reading file with the FTP adapter. As the file can be huge, we are using batching option with a batch size of 2000 records. The records in the file is in the Order and Lines format. In Bpel i am inserting those records in database. I have one Db adapter having two tables Orders and lines with a 1:M relation.
    Now in the positive test case scenario, Bpel is able to read the file , split it in to batches and successfully inserts the records in the tables (Orders and Lines).
    Now when i am testing a negative test case, i have violated the Primary key constraint in one of the record. Now ideally the batch which encountered the issue and failed, none of the records should get inserted in the Orders and Lines table, but that is not happening. In our case, no records are getting inserted to the lines table (which is as expected) but few records are getting inserted in to the Order table of the failed batch.
    F.Y.I : Number of Records in the batch can be less then Batch Size.
    Any clue on this?
    Thanks.

    Hi Jaap,
    thanks for your reply...
    Sorry for being a little unclear ...
    I have followed the instrunctions in the on-line help of OWB , and :
    1) have imported four code tables into a source module . The tables are in an Oracle RDBMS 10g .
    2)I have created manually a table - using the OWB client tool - which consisted of some of the columns found in the four tables .
    3)I have inserted a Join Operator in order to correlate the four table columns with the columns of the target module.
    4)I have defined the Join condition of the the four table columns. In my first post , unintenionally I have written that I removed the Join Operator. I meant the Join Condition in order to make sure that the join condition isn't the reason that the records are not inserted....
    5)I have validated and generated the mapping without any errors - not even warnings.
    6)I have selected Project->Deployment Manager and I selected the OWB_RUN user - which is the Runtime Repository owner.
    7)I have selected the objects to deploy - the mapping , the one table I created in the target module. The deployment has succedded , and the table was created but no records are found in it.
    Now , is there a possibility that this problem is caused by absence of the privileges to insert the rows...?
    How can I check that the source tables are filled on the registered location of your source module?
    Another basic question...
    I have created a Runtime Repository owner - called owb_run- using the Runtime Repository Assistant . I use this user in order to deploy the mapping and the new table. I want -and it is created- that table in another schema (user his_dw) - even another database in the same server. Should I run the Runtime Repository owner in this schema (his_dw)???
    Thanks , again a lot
    Simon

  • NEXT button on report portlet shows 10 rows-maximum rows per page=9999

    I created a report with the DB Provider portlet. I set the maximum rows per page to 9999 when displayed as a portlet and as a full page. I put the portlet on a page and it only shows 10 rows and then the next button.
    If there are exactly 10 rows I get the next button. If I click it the next page loads with no data. If there are less than 10 rows there is no button.
    If I create a new portlet it does not do this. I'm only having the problem with one portlet now but I've seen it in the past. I usually just build the portlet again but I would like to know how to fix it.
    I can't customize the portlet because of bug #3467032.

    Select customize and set the number of rows to 9999. I'm guessing that it's currently set to 10.

  • SQL merge and after insert or update on ... for each row fires too often?

    Hello,
    there is a base table, which has a companion history table
    - lets say USER_DATA & USER_DATA_HIST.
    For each update on USER_DATA there has to be recorded the old condition of the USER_DATA record into the USER_DATA_HIST (insert new record)
    - to have the history of changes to USER_DATA.
    The first approach was to do the insert for the row trigger:
    trigger user_data_tr_aiu after insert or update on user_data for each rowBut the performance was bad, because for a bulk update to USER_DATA, there have been individual inserts per records.
    So i tried a trick:
    Instead of doing the real insert into USER_DATA_HIST, i collect the USER_DATA_HIST data into a pl/sql collection first.
    And later i do a bulk insert for the collection in the USER_DATA_HIST table with stmt trigger:
    trigger user_data_tr_ra after insert or update on user_dataBut sometimes i recognize, that the list of entries saved in the pl/sql collection are more than my USER_DATA records being updated.
    (BTW, for the update i use SQL merge, because it's driven by another table.)
    As there is a uniq tracking_id in USER_DATA record, i could identify, that there are duplicates.
    If i sort for the tracking_id and remove duplicate i get exactly the #no of records updated by the SQL merge.
    So how comes, that there are duplicates?
    I can try to make a sample 'sqlplus' program, but it will take some time.
    But maybe somebody knows already about some issues here(?!)
    - many thanks!
    best regards,
    Frank

    Hello
    Not sure really. Although it shouldn't take long to do a test case - it only took me 10 mins....
    SQL>
    SQL> create table USER_DATA
      2  (   id      number,
      3      col1    varchar2(100)
      4  )
      5  /
    Table created.
    SQL>
    SQL> CREATE TABLE USER_DATA_HIST
      2  (   id      number,
      3      col1    varchar2(100),
      4      tmsp    timestamp
      5  )
      6  /
    Table created.
    SQL>
    SQL> CREATE OR REPLACE PACKAGE pkg_audit_user_data
      2  IS
      3
      4      PROCEDURE p_Init;
      5
      6      PROCEDURE p_Log
      7      (   air_UserData        IN user_data%ROWTYPE
      8      );
      9
    10      PROCEDURE p_Write;
    11  END;
    12  /
    Package created.
    SQL> CREATE OR REPLACE PACKAGE BODY pkg_audit_user_data
      2  IS
      3
      4      TYPE tt_UserData        IS TABLE OF user_data_hist%ROWTYPE INDEX BY BINARY_INTEGER;
      5
      6      pt_UserData             tt_UserData;
      7
      8      PROCEDURE p_Init
      9      IS
    10
    11      BEGIN
    12
    13
    14          IF pt_UserData.COUNT > 0 THEN
    15
    16              pt_UserData.DELETE;
    17
    18          END IF;
    19
    20      END;
    21
    22      PROCEDURE p_Log
    23      (   air_UserData        IN user_data%ROWTYPE
    24      )
    25      IS
    26          ln_Idx              BINARY_INTEGER;
    27
    28      BEGIN
    29
    30          ln_Idx := pt_UserData.COUNT + 1;
    31
    32          pt_UserData(ln_Idx).id     := air_UserData.id;
    33          pt_UserData(ln_Idx).col1   := air_UserData.col1;
    34          pt_UserData(ln_Idx).tmsp   := SYSTIMESTAMP;
    35
    36      END;
    37
    38      PROCEDURE p_Write
    39      IS
    40
    41      BEGIN
    42
    43          FORALL li_Idx IN INDICES OF pt_UserData
    44              INSERT
    45              INTO
    46                  user_data_hist
    47              VALUES
    48                  pt_UserData(li_Idx);
    49
    50      END;
    51  END;
    52  /
    Package body created.
    SQL>
    SQL> CREATE OR REPLACE TRIGGER preu_s_user_data BEFORE UPDATE ON user_data
      2  DECLARE
      3
      4  BEGIN
      5
      6      pkg_audit_user_data.p_Init;
      7
      8  END;
      9  /
    Trigger created.
    SQL> CREATE OR REPLACE TRIGGER preu_r_user_data BEFORE UPDATE ON user_data
      2  FOR EACH ROW
      3  DECLARE
      4
      5      lc_Row      user_data%ROWTYPE;
      6
      7  BEGIN
      8
      9      lc_Row.id   := :NEW.id;
    10      lc_Row.col1 := :NEW.col1;
    11
    12      pkg_audit_user_data.p_Log
    13      (   lc_Row
    14      );
    15
    16  END;
    17  /
    Trigger created.
    SQL> CREATE OR REPLACE TRIGGER postu_s_user_data AFTER UPDATE ON user_data
      2  DECLARE
      3
      4  BEGIN
      5
      6      pkg_audit_user_data.p_Write;
      7
      8  END;
      9  /
    Trigger created.
    SQL>
    SQL>
    SQL> insert
      2  into
      3      user_data
      4  select
      5      rownum,
      6      dbms_random.string('u',20)
      7  from
      8      dual
      9  connect by
    10      level <=10
    11  /
    10 rows created.
    SQL> select * from user_data
      2  /
            ID COL1
             1 GVZHKXSSJZHUSLLIDQTO
             2 QVNXLTGJXFUDUHGYKANI
             3 GTVHDCJAXLJFVTFSPFQI
             4 CNVEGOTDLZQJJPVUXWYJ
             5 FPOTZAWKMWHNOJMMIOKP
             6 BZKHAFATQDBUVFBCOSPT
             7 LAQAIDVREFJZWIQFUPMP
             8 DXFICIPCBCFTPAPKDGZF
             9 KKSMMRAQUORRPUBNJFCK
            10 GBLTFZJAOPKFZFCQPGYW
    10 rows selected.
    SQL> select * from user_data_hist
      2  /
    no rows selected
    SQL>
    SQL> MERGE
      2  INTO
      3      user_data a
      4  USING
      5  (   SELECT
      6          rownum + 8 id,
      7          dbms_random.string('u',20) col1
      8      FROM
      9          dual
    10      CONNECT BY
    11          level <= 10
    12  ) b
    13  ON (a.id = b.id)
    14  WHEN MATCHED THEN
    15      UPDATE SET a.col1 = b.col1
    16  WHEN NOT MATCHED THEN
    17      INSERT(a.id,a.col1)
    18      VALUES (b.id,b.col1)
    19  /
    10 rows merged.
    SQL> select * from user_data_hist
      2  /
            ID COL1                 TMSP
             9 XGURXHHZGSUKILYQKBNB 05-AUG-11 10.04.15.577989
            10 HLVUTUIFBAKGMXBDJTSL 05-AUG-11 10.04.15.578090
    SQL> select * from v$version
      2  /
    BANNER
    Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bi
    PL/SQL Release 10.2.0.4.0 - Production
    CORE    10.2.0.4.0      Production
    TNS for Linux: Version 10.2.0.4.0 - Production
    NLSRTL Version 10.2.0.4.0 - ProductionHTH
    David

  • OBIEE 11g: Report not displaying maximum rows per page

    I have an OBIEE report that is not displaying the maximum rows per page. When I click on the UP/DOWN arrow at the bottom of the report there is no change.

    Hi,
    Refer the below link.
    http://satyaobieesolutions.blogspot.in/2012/08/limit-row-in-table-and-graph-in-initial.html
    OR,
    https://supporthtml.oracle.com/ep/faces/secure/km/DocumentDisplay.jspx?id=1198961.1
    Add the following to your instanceconfig and then restart the presentation service.
    file path
    D:\Oracle\Middleware\instances\instance1\config\OracleBIPresentationServicesComponent\coreapplication_obips1\instanceconfig.xml just add below lines
    <Views>
    <Pivot>
    <MaxCells>6500000</MaxCells>
    <MaxVisibleColumns>100</MaxVisibleColumns>
    <MaxVisiblePages>100</MaxVisiblePages>
    <MaxVisibleRows>65000</MaxVisibleRows>
    <MaxVisibleSections>25</MaxVisibleSections>
    <DefaultRowsDisplayed>500</DefaultRowsDisplayed>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><DefaultRowsDisplayedInDelivery>75</DefaultRowsDisplayedInDelivery>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><DefaultRowsDisplayedInDownload>65000</DefaultRowsDisplayedInDownload>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><DisableAutoPreview>false</DisableAutoPreview>
    </Pivot>
    <Table>
    <MaxCells>6500000</MaxCells>
    <MaxVisiblePages>100</MaxVisiblePages>
    <MaxVisibleRows>65000</MaxVisibleRows>
    <MaxVisibleSections>25</MaxVisibleSections>
    <DefaultRowsDisplayed>500</DefaultRowsDisplayed>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><DefaultRowsDisplayedInDelivery>75</DefaultRowsDisplayedInDelivery>
    <!--This Configuration setting is managed by Oracle Business Intelligence Enterprise Manager--><DefaultRowsDisplayedInDownload>65000</DefaultRowsDisplayedInDownload>
    </Table>
    </Views>
    Hope this help's
    Thanks,
    Satya

  • Disabling the display maximum rows per page

    My project include many reports with table view, displaying 100 rows.
    There is always arrows icon with the option to display maximum rows per page.
    If user click on that option the OBIEE crashed, since there are a lot of rows.
    I cannot decrese this number since all rows are required for other charts views.
    a. Is there a way to set this parameter only for the table view?
    b. Is there a way to disable this icon or not presenting it to the user?
    c. Is there global way to coonfigure the default displayed rows number (e.g. from 100 to 20)?

    unfortunately, you are limited to the SRW built-ins in Reports as far as setting object properties is concerned. I have checked this with the Metalink staff and I did get an admission out of them that this is true.
    You can set the property in the property palette, but I'm not sure that this helps you as I take it you have a condition you would like to implement. I wish I could help you further, but Reports is limited in some areas, and setting object properties programmatically is one of them.

  • Page Break / Maximum Rows per Page

    Hi - One more question....
    I have a group left report and want to keep each master group (the value that is grouped on the left) on its own separate page. In Reports 6i I would have set the Max Rows Per Page property to 1 on the outer repeating frame and that would make sure I had at most 1 master record per page.
    The only way I can see to replicate that functionality in XML Pub is to insert a page break (split-by-page-break) before the end-for-each tag on the group. It looks good in PDF, but in Excel or HTML it repeats the full header with Logo and titles and looks pretty ridiculous.
    So the $1 Million dollar question is how can I make sure there is at most 1 master group per page, but when I output to Excel I only get the header once? Because if I take out the split-by-page-break I get the desired output in Excel but not in the PDF. I really want to be able to use 1 template for all output types because it seems silly to have a separate template depending on the output type.
    Ideas?

    PDF, understand what is page,
    HTML is a single page view and it does nto understand, and same is the case for Excel.

  • Changing Maximum Number of Rows Per Page to Include

    I am trying to change the "Maximum Number of Rows Per Page to Include" using Enterprise Manager > Coreapplication > Capacity Management > Performance.
    It was default to 75. I need to change it to around 5,000. I have a report which generates 5,000 records and I use Agent to send it to our user at a scheduled time.
    I change 75 to 5000, activate the changes and restart the BI. After restarting it automatically reverts back to 75.
    Has anyone experienced this?
    My instanceconfig has contains this:
    <Views>
    <Pivot>
    <!--This Configuration setting is managed by Oracle Enterprise Manager Fusion Middleware Control-->
    <DefaultRowsDisplayedInDelivery>75</DefaultRowsDisplayedInDelivery>
    <!--This Configuration setting is managed by Oracle Enterprise Manager Fusion Middleware Control-->
    <DefaultRowsDisplayedInDownload>5000000</DefaultRowsDisplayedInDownload>
    <!--This Configuration setting is managed by Oracle Enterprise Manager Fusion Middleware Control-->
    <DisableAutoPreview>false</DisableAutoPreview>
    <MaxCells>1920000</MaxCells>
    <MaxVisiblePages>2500</MaxVisiblePages>
    <MaxVisibleRows>1000000</MaxVisibleRows>
    <MaxVisibleSections>5000</MaxVisibleSections>
    </Pivot>
    <Table>
    <!--This Configuration setting is managed by Oracle Enterprise Manager Fusion Middleware Control-->
    <DefaultRowsDisplayedInDelivery>75</DefaultRowsDisplayedInDelivery>
    <!--This Configuration setting is managed by Oracle Enterprise Manager Fusion Middleware Control-->
    <DefaultRowsDisplayedInDownload>5000000</DefaultRowsDisplayedInDownload>
    <MaxCells>1920000</MaxCells>
    <MaxVisiblePages>2500</MaxVisiblePages>
    <MaxVisibleRows>1000000</MaxVisibleRows>
    <MaxVisibleSections>2000000</MaxVisibleSections>
    </Table>
    </Views>
    Any help would be greatly appreciated.
    Thanks!

    Hi,
    Editing in EM should change the settings. Stop all the services and restart
    In the individual reports for tables and pivot tables >> Edit the table properties and add the number of 'Rows Per Page'
    Hope this helped/ answered
    Regards
    MuRam

  • Need help with inserting rows in ResultSet and JTable

    hello Guru!
    i have inserted a row in my result set and i want that my table shows this row promptly after i have inserted it in my result set...
    but when i use following code for my resultset:
    rs.moveToInsertRow();
    rs.updateInt(1,nr);
    rs.updateString(2, name);
    rs.insertRow();
    Record are inserted in resultset and database but not shown in my JTable??
    Anyone a Clue to without reexecuting the query how can i display inserted row in JTable
    http://download-west.oracle.com/docs/cd/A87860_01/doc/java.817/a83724/resltse7.h
    I have refrered the following links but still clue less help Guruuuuuuu
    i m really in trobble??????

    i am just near by the Solution using the Database Metadata
    by couldn't get the ideaaaa
    ==================================================
    http://download-west.oracle.com/docs/cd/A87860_01/doc/java.817/a83724/resltse7.htm
    Seeing Database Changes Made Internally and Externally
    This section discusses the ability of a result set to see the following:
    its own changes (DELETE, UPDATE, or INSERT operations within the result set), referred to as internal changes
    changes made from elsewhere (either from your own transaction outside the result set, or from other committed transactions), referred to as external changes
    Near the end of the section is a summary table.
    Note:
    External changes are referred to as "other's changes" in the Sun Microsystems JDBC 2.0 specification.
    Seeing Internal Changes
    The ability of an updatable result set to see its own changes depends on both the result set type and the kind of change (UPDATE, DELETE, or INSERT). This is discussed at various points throughout the "Updating Result Sets" section beginning on , and is summarized as follows:
    Internal DELETE operations are visible for scrollable result sets (scroll-sensitive or scroll-insensitive), but are not visible for forward-only result sets.
    After you delete a row in a scrollable result set, the preceding row becomes the new current row, and subsequent row numbers are updated accordingly.
    Internal UPDATE operations are always visible, regardless of the result set type (forward-only, scroll-sensitive, or scroll-insensitive).
    Internal INSERT operations are never visible, regardless of the result set type (neither forward-only, scroll-sensitive, nor scroll-insensitive).
    An internal change being "visible" essentially means that a subsequent getXXX() call will see the data changed by a preceding updateXXX() call on the same data item.
    JDBC 2.0 DatabaseMetaData objects include the following methods to verify this. Each takes a result set type as input (ResultSet.TYPE_FORWARD_ONLY, ResultSet.TYPE_SCROLL_SENSITIVE, or ResultSet.TYPE_SCROLL_INSENSITIVE).
    boolean ownDeletesAreVisible(int) throws SQLException
    boolean ownUpdatesAreVisible(int) throws SQLException
    boolean ownInsertsAreVisible(int) throws SQLException
    Note:
    When you make an internal change that causes a trigger to execute, the trigger changes are effectively external changes. However, if the trigger affects data in the row you are updating, you will see those changes for any scrollable/updatable result set, because an implicit row refetch occurs after the update.
    Seeing External Changes
    Only a scroll-sensitive result set can see external changes to the underlying database, and it can only see the changes from external UPDATE operations. Changes from external DELETE or INSERT operations are never visible.
    Note:
    Any discussion of seeing changes from outside the enclosing transaction presumes the transaction itself has an isolation level setting that allows the changes to be visible.
    For implementation details of scroll-sensitive result sets, including exactly how and how soon external updates become visible, see "Oracle Implementation of Scroll-Sensitive Result Sets".
    JDBC 2.0 DatabaseMetaData objects include the following methods to verify this. Each takes a result set type as input (ResultSet.TYPE_FORWARD_ONLY, ResultSet.TYPE_SCROLL_SENSITIVE, or ResultSet.TYPE_SCROLL_INSENSITIVE).
    boolean othersDeletesAreVisible(int) throws SQLException
    boolean othersUpdatesAreVisible(int) throws SQLException
    boolean othersInsertsAreVisible(int) throws SQLException
    Note:
    Explicit use of the refreshRow() method, described in "Refetching Rows", is distinct from this discussion of visibility. For example, even though external updates are "invisible" to a scroll-insensitive result set, you can explicitly refetch rows in a scroll-insensitive/updatable result set and retrieve external changes that have been made. "Visibility" refers only to the fact that the scroll-insensitive/updatable result set would not see such changes automatically and implicitly.
    Visibility versus Detection of External Changes
    Regarding changes made to the underlying database by external sources, there are two similar but distinct concepts with respect to visibility of the changes from your local result set:
    visibility of changes
    detection of changes
    A change being "visible" means that when you look at a row in the result set, you can see new data values from changes made by external sources to the corresponding row in the database.
    A change being "detected", however, means that the result set is aware that this is a new value since the result set was first populated.
    With Oracle8i release 8.1.6 and higher, even when an Oracle result set sees new data (as with an external UPDATE in a scroll-sensitive result set), it has no awareness that this data has changed since the result set was populated. Such changes are not "detected".
    JDBC 2.0 DatabaseMetaData objects include the following methods to verify this. Each takes a result set type as input (ResultSet.TYPE_FORWARD_ONLY, ResultSet.TYPE_SCROLL_SENSITIVE, or ResultSet.TYPE_SCROLL_INSENSITIVE).
    boolean deletesAreDetected(int) throws SQLException
    boolean updatesAreDetected(int) throws SQLException
    boolean insertsAreDetected(int) throws SQLException
    It follows, then, that result set methods specified by JDBC 2.0 to detect changes--rowDeleted(), rowUpdated(), and rowInserted()--will always return false with the 8.1.6 Oracle JDBC drivers. There is no use in calling them.
    Summary of Visibility of Internal and External Changes
    Table 12-1 summarizes the discussion in the preceding sections regarding whether a result set object in the Oracle JDBC implementation can see changes made internally through the result set itself, and changes made externally to the underlying database from elsewhere in your transaction or from other committed transactions.
    Table 12-1 Visibility of Internal and External Changes for Oracle JDBC
    Result Set Type Can See Internal DELETE? Can See Internal UPDATE? Can See Internal INSERT? Can See External DELETE? Can See External UPDATE? Can See External INSERT?
    forward-only
    no
    yes
    no
    no
    no
    no
    scroll-sensitive
    yes
    yes
    no
    no
    yes
    no
    scroll-insensitive
    yes
    yes
    no
    no
    no
    no
    For implementation details of scroll-sensitive result sets, including exactly how and how soon external updates become visible, see "Oracle Implementation of Scroll-Sensitive Result Sets".
    Notes:
    Remember that explicit use of the refreshRow() method, described in "Refetching Rows", is distinct from the concept of "visibility" of external changes. This is discussed in "Seeing External Changes".
    Remember that even when external changes are "visible", as with UPDATE operations underlying a scroll-sensitive result set, they are not "detected". The result set rowDeleted(), rowUpdated(), and rowInserted() methods always return false. This is further discussed in "Visibility versus Detection of External Changes".
    Oracle Implementation of Scroll-Sensitive Result Sets
    The Oracle implementation of scroll-sensitive result sets involves the concept of a window, with a window size that is based on the fetch size. The window size affects how often rows are updated in the result set.
    Once you establish a current row by moving to a specified row (as described in "Positioning in a Scrollable Result Set"), the window consists of the N rows in the result set starting with that row, where N is the fetch size being used by the result set (see "Fetch Size"). Note that there is no current row, and therefore no window, when a result set is first created. The default position is before the first row, which is not a valid current row.
    As you move from row to row, the window remains unchanged as long as the current row stays within that window. However, once you move to a new current row outside the window, you redefine the window to be the N rows starting with the new current row.
    Whenever the window is redefined, the N rows in the database corresponding to the rows in the new window are automatically refetched through an implicit call to the refreshRow() method (described in "Refetching Rows"), thereby updating the data throughout the new window.
    So external updates are not instantaneously visible in a scroll-sensitive result set; they are only visible after the automatic refetches just described.
    For a sample application that demonstrates the functionality of a scroll-sensitive result set, see "Scroll-Sensitive Result Set--ResultSet5.java".
    Note:
    Because this kind of refetching is not a highly efficient or optimized methodology, there are significant performance concerns. Consider carefully before using scroll-sensitive result sets as currently implemented. There is also a significant tradeoff between sensitivity and performance. The most sensitive result set is one with a fetch size of 1, which would result in the new current row being refetched every time you move between rows. However, this would have a significant impact on the performance of your application.
    how can i implement this using
    JDBC 2.0 DatabaseMetaData objects include the following methods to verify this. Each takes a result set type as input (ResultSet.TYPE_FORWARD_ONLY, ResultSet.TYPE_SCROLL_SENSITIVE, or ResultSet.TYPE_SCROLL_INSENSITIVE).
    boolean deletesAreDetected(int) throws SQLException
    boolean updatesAreDetected(int) throws SQLException
    boolean insertsAreDetected(int) throws SQLException

Maybe you are looking for

  • Error message when opening a .cp file (version issue?)

    I am using Captivate 4.0 and exporting to a .swf file. My project partner is using Captivate 3.0 I sent her my Captivate file to use as a template for our project and when she tries to open it, she gets an error message: Adobe Captivate could not ope

  • Plz plz help me out. Characterset Mismatch Problem

    I am using Oracle 8.1.7(Japanese) on Win2k professional (Japanese) and developing a multilingual application. * I have some columns of NVARCHAR2 datatype. * NLS_Lang entry of registry is right now NA. * Database procedures When i use insert or any ot

  • Contact form not working

    down load to third party web page looks fine click send says successful but i don't recessive mail any help??

  • Business Area Trail Balance

    Dear Experts, I am facing problem in Trail Balance We have implemented 4 Business Area in our company code, company code wise trail balance is matching but BA wise its not matching. I created 1 adjustment account  &  executed F.5D & F.5E transaction.

  • Placed .tiff files in illustrator have box around them.

    I am wondering how to properly place .tiff files in illustrator, so that they can easily be set up for commercial printing. I have created a simple sell sheet for a product that has 3 placed tiff images in it. The images are circular and appear fine