Fetching 10 records at a time

Product.Category dimension has 4 child nodes Accessories,Bikes,Clothing n Components.My problem is when I have thousands of first level nodes my application takes a lot of time to load. Is there a way to fetch only say 100 records at a time? So then when
i click a next button i get the next 100
Eg:On the 1st click of a button I fetch 2 members
WITH MEMBER [Measures].[ChildrenCount] AS
[Product].[Category].CurrentMember.Children.Count
SELECT [Measures].[ChildrenCount] ON 1
,TopCount([Product].[Category].Members, 2) on 0
FROM [Adventure Works]
This fetches only Accessories. Is there a way the fetch the next two records Bikes n Clothing on  click.
Then Components on the next click. So on an so forth.

Hi Tsunade,
According to your description, there are thousands of members on your cube. It will take long time to retrieve all the member at a time, in order to improve the performance, you are looking for a function to fetch 10 records at a time, right? Based on my
research, there is no such a functionally to work around this requirement currently.
If you have any concern about this behavior, you can submit a feedback at
http://connect.microsoft.com/SQLServer/Feedback and hope it is resolved in the next release of service pack or product. Your feedback enables Microsoft to make software and services the best that they can be, Microsoft might consider to add this feature
in the following release after official confirmation.
Regards,
Charlie Liao
TechNet Community Support

Similar Messages

  • Fetching many records all at once is no faster than fetching one at a time

    Hello,
    I am having a problem getting NI-Scope to perform adequately for my application.  I am sorry for the long post, but I have been going around and around with an NI engineer through email and I need some other input.
    I have the following software and equipment:
    LabView 8.5
    NI-Scope 3.4
    PXI-1033 chassis
    PXI-5105 digitizer card
    DELL Latitude D830 notebook computer with 4 GB RAM.
    I tested the transfer speed of my connection to the PXI-1033 chassis using the niScope Stream to Memory Maximum Transfer Rate.vi found here:
    http://zone.ni.com/devzone/cda/epd/p/id/5273.  The result was 101 MB/s.
    I am trying to set up a system whereby I can press the start button and acquire short waveforms which are individually triggered.  I wish to acquire these individually triggered waveforms indefinitely.  Furthermore, I wish to maximize the rate at which the triggers occur.   In the limiting case where I acquire records of one sample, the record size in memory is 512 bytes (Using the formula to calculate 'Allocated Onboard Memory per Record' found in the NI PXI/PCI-5105 Specifications under the heading 'Waveform Specifications' pg. 16.).  The PXI-5105 trigger re-arms in about 2 microseconds (500kHz), so to trigger at that rate indefinetely I would need a transfer speed of at least 256 Mb/s.  So clearly, in this case the limiting factor for increasing the rate I trigger at and still be able to acquire indefinetely is the rate at which I transfer records from memory to my PC.
    To maximize my record transfer rate, I should transfer many records at once using the Multi Fetch VI, as opposed to the theoretically slower method of transferring one at a time.  To compare the rate that I can transfer records using a transfer all at once or one at a time method, I modified the niScope EX Timestamps.vi to allow me to choose between these transfer methods by changing the constant wired to the Fetch Number of Records property node to either -1 or 1 repectively.  I also added a loop that ensures that all records are acquired before I begin the transfer, so that acquisition and trigger rates do not interfere with measuring the record transfer rate.  This modified VI is attached to this post.
    I have the following results for acquiring 10k records.  My measurements are done using the Profile Performance and Memory Tool.
    I am using a 250kHz analog pulse source.
    Fetching 10000 records 1 record at a time the niScope Multi Fetch
    Cluster takes a total time of 1546.9 milliseconds or 155 microseconds
    per record.
    Fetching 10000 records at once the niScope Multi Fetch Cluster takes a
    total time of 1703.1 milliseconds or 170 microseconds per record.
    I have tried this for larger and smaller total number of records, and the transfer time per is always around 170 microseconds per record regardless if I transfer one at a time or all at once.  But with a 100MB/s link and 512 byte record size, the Fetch speed should approach 5 microseconds per record as you increase the number of records fetched at once.
    With this my application will be limited to a trigger rate of 5kHz for running indefinetely, and it should be capable of closer to a 200kHz trigger rate for extended periods of time.  I have a feeling that I am missing something simple or am just confused about how the Fetch functions should work. Please enlighten me.
    Attachments:
    Timestamps.vi ‏73 KB

    Hi ESD
    Your numbers for testing the PXI bandwidth look good.  A value of
    approximately 100MB/s is reasonable when pulling data accross the PXI
    bus continuously in larger chunks.  This may decrease a little when
    working with MXI in comparison to using an embedded PXI controller.  I
    expect you were using the streaming example "niScope Stream to Memory
    Maximum Transfer Rate.vi" found here: http://zone.ni.com/devzone/cda/epd/p/id/5273.
    Acquiring multiple triggered records is a little different.  There are
    a few techniques that will help to make sure that you are able to fetch
    your data fast enough to be able to keep up with the acquired data or
    desired reference trigger rate.  You are certainly correct that it is
    more efficient to transfer larger amounts of data at once, instead of
    small amounts of data more frequently as the overhead due to DMA
    transfers becomes significant.
    The trend you saw that fetching less records was more efficient sounded odd.  So I ran your example and tracked down what was causing that trend.  I believe it is actually the for loop that you had in your acquisition loop.  I made a few modifications to the application to display the total fetch time to acquire 10000 records.  The best fetch time is when all records are pulled in at once. I left your code in the application but temporarily disabled the for loop to show the fetch performance. I also added a loop to ramp the fetch number up and graph the fetch times.  I will attach the modified application as well as the fetch results I saw on my system for reference.  When the for loop is enabled the performance was worst at 1 record fetches, The fetch time dipped  around the 500 records/fetch and began to ramp up again as the records/fetch increases to 10000.
    Note I am using the 2D I16 fetch as it is more efficient to keep the data unscaled.  I have also added an option to use immediate triggering - this is just because I was not near my hardware to physically connect a signal so I used the trigger holdoff property to simulate a given trigger rate.
    Hope this helps.  I was working in LabVIEW 8.5, if you are working with an earlier version let me know.
    Message Edited by Jennifer O on 04-12-2008 09:30 PM
    Attachments:
    RecordFetchingTest.vi ‏143 KB
    FetchTrend.JPG ‏37 KB

  • How to fetch some certain records at a time

    Hi everyone,
    I dont know how to fetch some records at a time,
    for some reason,there is not enouht memory.so I cant
    fethch all the records at a time.
    example, every time I only want to fetch 10
    records. Does anyone know that?
    thanks
    Krussi

    I think you may be getting setFetchSize(int) and setMaxRows(int) confused. I may have added to the confusion with my breaking out of the loop comment.
    setFetchSize(int) gives the driver a hint as to how many rows it should store in memory at a time internally, ie. within the driver. It has no real effect on how you write your program at all. Its only purpose is to limit memory usage by the driver. To limit the rows returned, use setMaxRows.
    For example, assume a result set has 20 rows and you use:
    rs.setFetchSize(5);
    while (rs.next()) {
       System.out.println("Rows: " + (++count));
    }  You still get a count from 1 to 20. However, internally the driver is only holding in memory 5 records at a time, after the 5 returned it will discard those and get the next 5. So instead of holding 20 rows, its only holding 5 at a time. If you change rs.setFetchSize(5) to rs.setMaxRows(5) you will only get a count from 1 to 5, the rest are silently discared and rs.next() returns false.
    Check your documentation to see if your driver supports the method. I believe many drivers don't. If not, I believe there is no way around multiple queries. Even if there is no rowid, as long as you order your queries you should be able to start again where you left off.
    Disclaimer: Both drivers I commonly use do not implement setFetchSize(int). Someone let me know if I've fudged something here.
    Good Luck

  • Time out error while fetching records from table BKPF

    Hi,
    I am fetching records from table BKPF using BUKRS & AWKEY in where clause. Query is as follows:
        SELECT BELNR  XBLNR  AWKEY
        FROM   BKPF
        INTO TABLE L_I_BKPF_TEMP
        PACKAGE SIZE 500
        WHERE BUKRS LIKE L_C_EG
        AND   AWKEY IN L_R_AWKEY .
          APPEND LINES OF L_I_BKPF_TEMP TO I_BKPF .
        ENDSELECT .
    Program is giving time out error. There are 25628 records in range L_R_AWKEY , i m fetching 500 records at a time using  PACKAGE SIZE. But the execution of prog stops on this query.
    Please suggest something to overcome this problem.

    Hi
    Rui is right,
    if you need to get the data by  operation parameters u have to use the fields AWTYP and AWKEY.
    In this selection u can omit the company code.
    SELECT BELNR XBLNR AWKEY FROM BKPF
           INTO TABLE L_I_BKPF_TEMP
                 PACKAGE SIZE 500
                      WHERE   AWTYP = <......> "<------------
                             AND AWKEY IN L_R_AWKEY .
           APPEND LINES OF L_I_BKPF_TEMP TO I_BKPF .
    ENDSELECT .
    Max

  • How can i fetch records from 3 tables in a single query  without using join

    Hi.
    Can any body please tell me <b>How can i fetch records from 3 tables with a single query  without using joins</b>
    Thanx
    prabhudutta

    Hi Prabgudutta,
    We can fetch the data by using the views concept.
    Go throuth this info we can know the how to create view and same like database table only we can fetch the data.
    Views conatin the data at runtime only.
    Four different view types are supported. These differ in the
    way in which the view is implemented and in the methods
    permitted for accessing the view data.
    Database views are implemented with an equivalent view on
    the database.
    Projection views are used to hide fields of a table (only
    projection).
    Help views can be used as selection method in search helps.
    Maintenance views permit you to maintain the data
    distributed
    on several tables for one application object at one time.
    step by step creation of Maintenance view:
    With the help of the table maintenance generator, you are able to maintain the ENTRIES of the table in SM30 transaction.
    It can be set in transaction SE11 - Tools - Table maintenance generator.
    Table maintanance Generator is used to manually input values using transaction sm30
    follow below steps
    1) go to se11 check table maintanance check box under attributes tab
    2) utilities-table maintanance Generator-> create function group and assign it under
    function group input box. Also assign authorization group default &NC& .
    3) select standard recording routine radio in table table mainitainence generator to move table
    contents to quality and production by assigning it to request.
    4) select maintaience type as single step.
    5) maintainence screen as system generated numbers this dialog box appears when you click on create button
    6) save and activate table
    One step, two step in Table Maintenance Generator
    Single step: Only overview screen is created i.e. the Table Maintenance Program will have only one screen where you can add, delete or edit records.
    Two step: Two screens namely the overview screen and Single screen are created. The user can see the key fields in the first screen and can further go on to edit further details.
    SM30 is used for table maintenance(addition or deletion of records),
    For all the tables in SE11 for which Table maintenance is selected , they can be maintained in SM30
    Sm30 is used to maintain the table ,i.e to delete ,insert or modify the field values and all..
    It creates the maintenance screen for u for the aprticular table as the maintenance is not allowed for the table..
    In the SE11 delivery and maintenance tab, keep the maintenance allowed..
    Then come to the SM30 and then enter the table name and press maintain..,
    Give the authorization group if necessary and give the function group and then select maintenance type as one step and give the screen numbers as system specified..
    Then create,,,
    Then u will able to see the maintenance view for the table in which u can able to insert and delete the table values...
    We use SM30 transaction for entering values into any DB table.
    First we create a table in SE11 and create the table maintenance generator for that Table using (utilities-> table maintenance generator) and create it.
    Then it will create a View.
    After that from SM30, enter the table name and Maintain, create new entries, change the existing entries for that table.
    Hope this resolves your query.
    Reward all the helpful answers.
    Rgds,
    P.Naganjana Reddy

  • How to fetch records with two recordset

    i want fetch records from the table using two recordset like
    ResultSet r1=stmt.executeQuery("select * from pf_data");
    i want to take another recodeset
    ResultSet r2=stmt.executeQuery("select * from finyr");
    I am using Microsoft SQL server as my database.
    I am also closing the recordset like r1.close();
    then i cannot take two recordset at the same time.

    Hi,
    This is an example of a Top-N problem, and here's one way to do it:
    WITH got_r_num AS
        SELECT  item_id, supp_id, rnk                         -- RANK is not a good column name
        ,       DENSE_RANK () OVER ( PARTITION BY  item_id    -- Just guessing
                                     ORDER BY      rnk
                                   )  AS r_num
        FROM    table_x
        WHERE   item_id  = 2      -- If desired
    SELECT    item_id, supp_id. rnk
    FROM      got_r_num
    WHERE     r_num    <= 2
    RANK is the name of a built-in function in Oracle, so it will cause lots of trouble if you use it as a column name.  (The RANK function is closely related to DENSE_RANK, used above.)
    I hope this answers your question.
    If not, post  a little sample data (CREATE TABLE and INSERT statements, relevant columns only) for all tables involved, and also post the results you want from that data.
    Point out where the query above is giving the wrong results, and explain, using specific examples, how you get the correct results from the given data in those places.  If you changed the query at all, post your code.
    Always say which version of Oracle you're using (e.g., 11.2.0.2.0).
    See the forum FAQ: https://forums.oracle.com/message/9362002

  • Database adapter for performing sync read of N records at a time

    Would anyone have an example and/or documentation related to performing a select operation in a database adapter where the adapter can fetch N number of records at a time? The database adapter would not be an activation agent/point polling for changes to any records, but invoked in the middle of the process. I searched for this in the samples folder and online but found nothing.
    Any help would be appreciated.
    Thanks.

    Hi,
    Do you want to know the best way to integrate Oracle BPEL to OTM? Then it's simple webservice call from BPEL using http binding.
    If you want to connect to Apps inbound then you can either use DB adapter or JMS adapter or Apps adapter which ever best suites your business requirement.
    Rgds,
    Mandrita.

  • Fetching records in block

    hi guruz,
    i have requirement something like that in a table more than 3 lakh record are there.
    i want to fetch records in a block of 1000 . how can i achive this.
    i already implemented one way that fetch all records (3 lakhs) one time , and process those in internal table .
    But selecting 3lakhs records at one time abap application server giving timeout error.
    some cofigraution required or what.
    please guide me.
    warm regards.

    Hi,
    what about doing it package by package ( 1 package = 1000 rows) and process the int. table as you like:
    DATA: itab TYPE STANDARD TABLE OF SCARR WITH NON-UNIQUE
                     DEFAULT KEY INITIAL SIZE 10.
    FIELD-SYMBOLS: <FS> TYPE scarr.
    SELECT * INTO TABLE itab PACKAGE SIZE 1000 FROM scarr.
      LOOP AT itab ASSIGNING <FS>.
                ... process it
      ENDLOOP.
    ENDSELECT.
    Set the package size high enough that you have not many additional selects
    (10.000 or 50.000 should be also ok).
    Please be aware that the int. table is only defined inside the SELECT / ENDSELECT.
    bye
    yk

  • Showing 10 records at a time in a multirecord block

    Hello,
    I have a requirement to show 10 records at a time,consider a table in which
    there are 100 records which is associated to a multirecord block,in order to have a web look and feel I need to show first 10(1 to 10) record on load,when the user presses the button "Next 10 records", I need to replace the first 10 records with new set of 10 records(11 - 20),
    Please note that no of records to be displayed at a time will be 10 and our requirement is avoid scroll bar.
    Any suggestions as to how I need to accomplish the task above.
    Thanks and Regards
    Mohan

    You don't need a special button for this because the user can already navigate using the cursor keys and page up/down keys.
    For cursor keys use the key-up, key-down, key-nxtrec and key-prvrec triggers.
    For page up/down use the key-scrup and key-scrdown triggers.
    I suppose the menu option "Query > Fetch Next Set" (key-nxtset) should operate the same as page down.
    An example of the key-scrdown trigger is thisdeclare
         numrecs integer := get_block_property('EMP',records_displayed);
         toprec integer := get_block_property('EMP',top_record);
         thisrec integer := :system.trigger_record;
    begin
         go_record(toprec + (2 * numrecs) - 1);
         go_record(thisrec + numrecs);
    end;This will advance by as many records are in the block and leave the cursor positioned on the same record relative to the top of the block. Eg if there are 10 rows in the block and you're on the first set then the top record will be 1 and you want to make the new top record 11. So you need to call go_record(20), which makes that the bottom record and 11 the top one.
    You'll need to work out how the form should behave when the number of records in the block is not exactly divisible by the number of records in the data because forms won't allow blank records to pad out the bottom of the block and records_displayed isn't a settable property at runtime. So if there are 10 rows in the block and 103 rows of data then the penultimate page will show records 91-100 and the last page will show records 94-103. Should scrolling back up show record 91 at the top, or 84?
    For the key-down/key-nxtrec trigger you'll probably want something like thisdeclare
         numrecs integer := get_block_property('EMP',records_displayed);
    begin
         if :system.last_record != 'TRUE' then
              if mod(:system.cursor_record,numrecs) = 0 then
                   -- last record in this set
                   go_record(:system.cursor_record + numrecs);
                   go_record(:system.cursor_record - (numrecs - 1));
              else
                   next_record;
              end if;
         end if;
    end;If you're at the bottom of one set it will show the next set and go to the top record, otherwise it will just go to the next record. But again, this breaks down when the number of rows in the block and in the data don't divide up evenly.
    I hope this is enough to get you started.

  • Fetch records having timing before 3 PM

    Hellow All,
    We are using Oracle 10G R2 on windows.
    I have a requirment that i need to fetch the records which are having cut off time before 3 pm.
    query should not fetch records after 3 pm.
    Please help how we can achive this.
    For testing purpose I have records like below
    SQL> select empno,emptype,to_char(jdate,'dd-mm-yyyy hh:mi:ss')jdate from test;
         EMPNO EMPTYPE              JDATE
             6 back office          24-12-0012 03:00:00
             1 HR                   01-01-2012 03:01:00
             2 Admin                01-12-2011 03:20:00
             3 MARKETING            01-01-2012 03:00:00
             4 FACILITY             12-02-2012 03:00:00
             8 IT                   10-06-2010 03:14:00
             9 Finance              07-11-2012 03:10:00
            10 "100HR"              07-11-2012 03:31:00
    8 rows selected.In the above case records having time more than 3 pm should not fetch.

    I have tried below query, it is giving me
    /* Formatted on 1/21/2013 3:25:20 PM (QP5 v5.215.12089.38647) */
    WITH test
         AS (SELECT 6 empno,
                    'back office' emptype,
                    TO_DATE ('24-12-2012 15:00:00', 'dd-mm-yyyy hh24:mi:ss')
                       jdate
               FROM DUAL
             UNION ALL
             SELECT 6 empno,
                    'back office' emptype,
                    TO_DATE ('24-12-2012 15:40:00', 'dd-mm-yyyy hh24:mi:ss')
                       jdate
               FROM DUAL
             UNION ALL
             SELECT 1,
                    'HR',
                    TO_DATE ('01-01-2012 03:01:00', 'dd-mm-yyyy hh24:mi:ss')
               FROM DUAL
             UNION ALL
             SELECT 2,
                    'Admin',
                    TO_DATE ('01-12-2011 03:20:00', 'dd-mm-yyyy hh24:mi:ss')
               FROM DUAL
             UNION ALL
             SELECT 3,
                    'MARKETING',
                    TO_DATE ('01-01-2012 03:00:00', 'dd-mm-yyyy hh24:mi:ss')
               FROM DUAL
             UNION ALL
             SELECT 4,
                    'FACILITY',
                    TO_DATE ('12-02-2012 03:00:00', 'dd-mm-yyyy hh24:mi:ss')
               FROM DUAL
             UNION ALL
             SELECT 8,
                    'IT',
                    TO_DATE ('10-06-2010 03:14:00', 'dd-mm-yyyy hh24:mi:ss')
               FROM DUAL
             UNION ALL
             SELECT 9,
                    'Finance',
                    TO_DATE ('07-11-2012 03:10:00', 'dd-mm-yyyy hh24:mi:ss')
               FROM DUAL
             UNION ALL
             SELECT 10,
                    '100HR',
                    TO_DATE ('07-11-2012 03:31:00', 'dd-mm-yyyy hh24:mi:ss')
               FROM DUAL)
    SELECT empno, emptype, TO_CHAR (jdate, 'dd-mm-yyyy hh24:mi:ss') jdate
      FROM test
    WHERE TO_CHAR (jdate, 'hh:mi:ss') <= '03:00:00'records for 6, 3 and 4 empno's check if this is your requirement

  • Getting virtual memory error when fetching records from database

    HI,
    I am using Oracle as Database and the Oracle JDBC driver. I have simple code which is a select statement but the problem is the resultset dies while fetching the data i.e. 5,50,000. And it gives me the memory error as its storing all in the memory. One of the way which i have serched in the old threads is using the batch method fetching rows at a time but can you tell me how to implement in my code. I am pasting my code.
    The overall functionality of my code is that it's reterving data from database and generating an XML file that would be validated with a DTD.
    //My Code
    public class Invoicef3 implements ExtractF3 {
         final String queryString = "select * from hsbc_f3_statement
    order by bill_no, duplicate,
    invoice_address1,
    invoice_address2,
    invoice_address3,
    invoice_address4,
    invoice_address5,
    invoice_address6,
    main_section, order_1, page,
    section, product_category,
    sub_sect_1, order_2,
    sub_sect_2, child_product,
    sub_sect_3, account,
    line,entry_date, currency, tier";
         public ArrayList process() {
              Connection con = null;
              Statement stmt = null;
              ResultSet rset = null;
              ArrayList arr1 = null;
              try {
                   con =
    ConnectionManager.getConnection();
                   stmt = con.createStatement();
              rset = stmt.executeQuery(queryString);
                   arr1 = new ArrayList();
                   while (rset.next()) {
                        arr1.add(
                             new F3StatementExtract(
                                  rset.getString(1),
                                  rset.getInt(2),
                                  rset.getString(3),
                                  rset.getInt(4),
                                  rset.getInt(5),
                                  rset.getString(6),
                                  rset.getInt(7),
                                  rset.getString(8),
                                  rset.getInt(9),
                                  rset.getString(10),
                                  rset.getInt(11)));
                   rset.close();
                   stmt.close();
              } catch (SQLException e) {
                   e.printStackTrace();
              } finally {
                   ConnectionManager.close(rset);
                   ConnectionManager.close(stmt);
                   ConnectionManager.close(con);
              return arr1;
    }

    The problem is that you are fetching and processing all the rows for the query, which the VM cannot handle given the heap space available. The points you could think over are:
    * Allocate more heap memory (this would help only to a limited extent)
    * Try to process only a few records at a time instead of all of them (there is actually no need to process all the records at a time. Try processing records in lots of say 1000)
    * Avoid selecting all the columns [SELECT *] from the table, if all of them are not going to be used.
    There is a slight change i have done in the code is that i am using two quereies now one is fetching all the Bills and the secondquery is fetching all the data for the relevant BILL.
    //My Code
    public class Invoicef3 implements ExtractF3 {
         /*Query to get distinct bill numbers*/
         final String queryString1 =
              "select distinct(bill_no) from hsbc_print_bills";
         /*Query to get distinct bill numbers statement details*/
         final String queryString =
              "select * from hsbc_f3_statement where bill_no='";
         public ArrayList process() {
              Connection con = null;
              Statement stmt = null;
              ResultSet rset = null;
              ArrayList arr1 = null;
              ArrayList arr2 = null;
              try {
                   con = ConnectionManager.getConnection();
                   stmt = con.createStatement();
                   rset = stmt.executeQuery(queryString1);
                   arr1 = new ArrayList();
                   while (rset.next()) {
                        arr1.add(new F3BillExtract(rset.getString(1))); //generating the Bill_No's
                   System.out.print(arr1.size());
                   rset.close();
                   stmt.close();
                   for (int i = 0; i < arr1.size(); i++) {
                        stmt = con.createStatement();
                        rset =
                             stmt.executeQuery(
                                  queryString
                                       + (((F3BillExtract) arr1.get(i)).getBill_No())
                                       + "'");
                        arr2 = new ArrayList();
                        /*Fetching the statement Details of the particular Bill_No*/
                        while (rset.next()) {
                             arr2.add(
                                  new F3StatementExtract(
                                       rset.getString(1),
                                       rset.getInt(2),
                                       rset.getString(3),
                                       rset.getInt(4),
                                       rset.getInt(5),
                                       rset.getString(6),
                                       rset.getInt(7),
                                       rset.getString(8),
                                       rset.getInt(9),
                                       rset.getString(10),
                                       rset.getInt(11),
                                       rset.getString(12),
                                       rset.getFloat(13),
                                       rset.getDate(14),
                                       rset.getString(15),
                                       rset.getInt(16),
                                       rset.getString(17),
                                       rset.getString(18),
                                       rset.getString(19),
                                       rset.getString(20),
                                       rset.getString(21),
                                       rset.getString(22),
                                       rset.getString(23),
                                       rset.getString(24),
                                       rset.getString(25),
                                       rset.getString(26),
                                       rset.getString(27),
                                       rset.getString(28),
                                       rset.getString(29),
                                       rset.getString(30),
                                       rset.getDate(31),
                                       rset.getDate(32),
                                       rset.getDate(33),
                                       rset.getDate(34),
                                       rset.getString(35),
                                       rset.getString(36),
                                       rset.getString(37),
                                       rset.getString(38),
                                       rset.getString(39),
                                       rset.getString(40),
                                       rset.getFloat(41),
                                       rset.getFloat(42),
                                       rset.getFloat(43),
                                       rset.getInt(44),
                                       rset.getFloat(45),
                                       rset.getString(46),
                                       rset.getString(47)));
                        rset.close();
                        stmt.close();
                        ((F3BillExtract) arr1.get(i)).setArr(arr2);
              } catch (SQLException e) {
                   e.printStackTrace();
              } finally {
                   ConnectionManager.close(rset);
                   ConnectionManager.close(stmt);
                   ConnectionManager.close(con);
              return arr1;
    }

  • Displaying one record at a time

    Is any body able to help me? I need to display one record at a time from a Select statement.
    Could I copy the results of the Select query from a table into an array for later display or is there a better alternative.
    Please, any part knowledge with regard to the above would be usefull.
    Regards
    Eddie

    What application do you want to use to display the data?
    You can use "rownum" in your SQL statement to ensure that you only get record N of a query. Asktom has a good article on this sort of windowing
    http://asktom.oracle.com/pls/ask/f?p=4950:8:::::F4950_P8_DISPLAYID:127412348064
    More commonly, however, you'll write a SQL query that returns all the rows, fetch a small number, display those to the user, and then fetch more records.
    Justin

  • Master_detail for more than one record at a time

    Hi,
    How can i display master_detail records for more than one records at a time, for example, i have two tables A and B , A has username and role and B has username and profile. here i wanted to display 10 users at a time on my 6i form with username, role and profile.
    i have created a master-detail relation ship with these tables when i'm executing F8 on blcok A , it displays 10 records on BlockA but, only one at a time on block B, how can i display all corresponding records on block B at a time.
    Thanks for your help.Bcj

    Thanks Roberts, that was realy informative due to some doubts i would like to confirm my requirements , i have two blocks A and B and each master record has only one detail record. but i wanted to display at least 10 master_detail relationships(records) on the form at a time, i would like to know is it possible to do without creating any table or view for example,
    data in table A,
    username role
    AAA R1
    BBB R2
    CCC R3
    data in table B,
    username profile
    AAA P1
    BBB P2
    CCC P3
    i wanted to display it on form like below,
    username role profile
    AAA R1 P1
    BBB R2 P2
    CCC R3 P3
    Also would like to know that how can i select data from dba_users, any restriction is there on forms 6i, i can select it on sqlplus.
    Thanks Again, Bcj

  • PDE-PLI031 Unable to fetch record from table tool_modulre

    Dear ALL
    I am creating PL/SQL Libraries in report builder.
    but When I try to save the Library to database, a error
    PDE-PLI031 Unable to fetch record from table tool_modulre.
    Would you please tell me how to solve this problem
    and why it coming
    thankyou very much
    pritam singh

    Hi ,
    Saving a library (.pll) to database would store the object inside specific tables that are to be created.
    If you are using 6i, then you should find toolbild & toolgrnt.sql files which you have to run in the order specified. The above scripts creates the necessary tables and henceforth you won't get those errors while saving.
    Hope this helps.
    Thanks,
    Vinod.

  • ? - Is there a way to validate 1 XML record at a time, using the SAX or oth

    Hello!
    Before running into space problems, i generated an XML file from a 'pipe delimited' file and and then processed that XML file thru a SAXParser 'validator' and the data was correctly validated, using the RELAXNG Schema patterns as the validation criteria!
    But as feared, the XML file was huge! (12 billion XML recs. generated from 1 billion 'pipe' recs.) and i am now trying to find a way to process 1 'pipe' record at a time, (ie) read 1 record from the 'pipe delimited' file, convert that rec. to an XML rec. and then send that 1 XML rec. thru the SAXParser 'validator', avoiding the build of a huge temporary XML file!
    After testing this approach, its looks like the SAXParser 'validator' (sp.parse) is expecting only (1) StringBufferInputStream as input,and after opening, reading and closing just (1) of the returned StringBufferInputStream objects, the validator wants to close up operations!
    Where i have the "<<<<<" you can see where i'm calling the the object.method that creates the 'pipe>XML' records 'sb.createxml' and will be returning many occurances of the StringBufferInputStream object, where (1) StringBufferInputStream object represents (1) 'pipe>XML' record!
    So what i'm wondering, is if there is a form of 'inputStream' class that can be loaded and processed at the same time! ie instead of requiring that the 'inputStream' object be loaded in it's entirety, before going to validation?
    Or if there is another XML 'validator' that can validate 1 XML record at a time, without requiring that the entire XML file be built first?
    1. ---------------------------------class: (SX2) ---------------------------------------------------------------------------------------------------------
    import ............
    public class SX2
    public static void main(String[] args) throws Exception
    MyDefaultHandler dh = new MyDefaultHandler();
    SX1 sx = new SX1();
    SAXParser sp = sx.getParser(args[0]);
    stbuf1 sb = new stbuf1();
    sp.parse(sb.createxml(args[1]),dh); <<<<<< createxml( ) see <<<<<<< below
    class MyDefaultHandler extends DefaultHandler {
    public int errcnt;
    "SX2.java" 87 lines, 2563 characters
    2. ----------------------------------class: (stbuf1) method: (createxml) ----------------------------------------------------------------------------
    public stbuf1 () { }
    public StringBufferInputStream createxml( String inputFile ) <<<<<< createxml(
    BufferedReader textReader = null;
    if ( (inputFile == null) || (inputFile.length() <= 1) )
    {     throw new NullPointerException("Delimiter Input File does not exist");
    String ele = new String();
    try {
    ele = new String();
    textReader = new BufferedReader(new FileReader(inputFile));
    String line = null; String SEPARATOR = "\\|"; String sToken = null;
    String hdr1=("<?xml version=#1.0# encoding=#UTF-8#?>"); hdr1=hdr1.replace('#','"');
    String hdr2=("<hlp_data>");
    String hdr3=("</hlp_data>");
    String hdr4=("<"+TABLE_NAME+">");
    String hdr5=("</"+TABLE_NAME+">");
    while ( (line = textReader.readLine()) != null )
    String[] sa = line.split(SEPARATOR);
    String elel = new String();
    for (int i = 0; i < NUM_COLS; i++)
    if (i>(sa.length-1)) { sToken = new String(); } else { sToken = sa; }
    elel="<"+_columnNames[i]+">"+sToken+"</"+_columnNames[i]+">";
    if (i==0) {
    ele=ele.concat(hdr1);ele=ele.concat(hdr2);ele=ele.concat(hdr4);ele=ele.concat(elel);
    else
    if (i==NUM_COLS - 1) {
    ele=ele.concat(elel);ele=ele.concat(hdr5);ele=ele.concat(hdr3);
    else {
    ele=ele.concat(elel);
    textReader.close();
    catch (IOException e) {
    return (new StringBufferInputStream(ele));
    public static void main( String args[] ) {
    stbuf1 genxml_obj = new stbuf1 ();
    String ptxt=new String(args[0]);
    genxml_obj.createxml(ptxt); }}

    Well,i think you can use the streaming API for xml processing provided by weblogic.It is pull model,not push model like SAX.with it,you can select the events you want without having to react to every event,and you can filter the events out.
    Sun also provide such streaming API for xml processing,and i got an very simple introduction about it on the Chinese Sun developer site.but i couldn't find any other infomation about it elsewhere! If you have such materials,please send to my email:[email protected],and if I have it,i will be sure to post the links here.hope it helps more or less:)
    @smile@

Maybe you are looking for

  • Windows xp pro stalls - blank screen- hardware or software problem?

    Hi Guys, My system is about one month old and has been awesome in performance and graphics. Suddenly it (windows xp pro) stalls after flashing the windows loading screen and it goes blank. Looks like a software problem? I am posting on this forum bec

  • List of Unfixed planned orders

    Hi - Is there a report in R/3 which lists all the existing unfixed planned orders in the system? Appreciate your help. Thanks Varun

  • How to divide resultset of a query in different ref cursors in a package

    Hi Oracle Gurus, I need to create a package which counts the no of rows returned by a select query on multiple tables and according to the returned rowcount inputs the resultset in a ref cursor. Procedure will be called by a .NET program and output p

  • Download Error in my Creative Cloud Apps Folder

    In my Adobe Creative Cloud "Apps" folder I'm getting the following message: Download error. Unable to reach Adobe servers. Please check your internet connectivity, firewall settings, and then try again. (400) Any ideas on what's going on?

  • Flash variables will not pass on enterFrame

    I currently have the following code setup to pass variables to a flash when a button is clicked: on click      flashSprite = sprite("videoplayer")      rootRef = flashSprite.getVariable("root1", false)      rootRef.setUpDoc(the moviePath&"xml"&dirPat