Best way to generate report on huge data table

Hi,
I am using Oracle11g.
I want to generate reports on transaction tables containing huge amount of data, on which very frequently DMLs are performing in real time.
i want to keep my report/result in RESULT_CACHE for 15 mins. active. whenever any insert/update runs on main tables the RESULT_CACHE is getting invalidated
my question is can i control/stop RESULT_CACHE relies on(table_name) invalidating, as Oracle11g invalidating RESULT_CACHE automatically.
my requirement is to not hit the main table again&again.
pls help..
Thanks in advance.
Vinod

910575 wrote:
Hi,
I am using Oracle11g.
I want to generate reports on transaction tables containing huge amount of data, on which very frequently DMLs are performing in real time.
i want to keep my report/result in RESULT_CACHE for 15 mins. active. whenever any insert/update runs on main tables the RESULT_CACHE is getting invalidated
my question is can i control/stop RESULT_CACHE relies on(table_name) invalidating, as Oracle11g invalidating RESULT_CACHE automatically.
my requirement is to not hit the main table again&again.
pls help..
It sounds as if you're trying to avoid contention on a very busy large table while users are experimenting with relatively small fractions of the total data set. The type of thing you're doing is probably about the best approach - though it sounds as if you are not using global temporary tables which could save you a bit of time and contention when refreshing each private data set.
Ideally, though, you probably want a front end tool that does client-side caching - i.e. pulls the data into the front-end tool and lets the user rearrange it cosmetically there until the user explicitly requests a new trip to the database. I think Oracle Discoverer has (had) some capability in this area. What's the scale of the work the users are doing - can you give us a few scenarios about how much raw data they will extract and what they want to do with it before they refresh it ?
Regards
Jonathan Lewis
http://jonathanlewis.wordpress.com
Author: <b><em>Oracle Core</em></b>
P.S. Oracle does have client-side caching technology - but your ability to use it is dependent on the tools you use. You might want to go over to one of the developer or BI forums to see what they say about this problem; they may give you a different perspective on it.
Edited by: Jonathan Lewis on Jan 31, 2012 6:55 PM

Similar Messages

  • Best way to produce reports if the data model is in RDBMS

    we have database available in RDBMS.we will have to create reports out of that database.
    Anyone suggest me best way(by considering all the factors) to get the reports if database is in RDBMS.
    what are all need to follow in physical layer and what are all need to follow in BMM layer.

    Thanks saichand for the response.
    Ex: lets say,if there are 25 tables in the database(in RDBMS,there are lot of joins exists between the tables like A->B->C->D->E->F->G->H, in this case if report is from A,E and G then which way i will have to create view in the physical layer)
    i have arrraged tables and columns below for better understanding
    A B c D E F G
    A1(pk) B1(PK) C1(PK) D1(PK) E1(PK) F1(PK) G1(PK)
    A2 A1(fk) B1(fk) C1(fk) D1(fk) E1(fk) F1(fk)
    A3 B2 C2 D2 E2 F2 G2
    from the above tables and columns,if the report should be with A1,E1,E3,G1,G2 fields in that case how to create view in the physical layer.
    for each and every report do we need to create seperate view?
    can you please elaborate the procedudure in different situations?
    Thanks in Advance.
    Edited by: user12077461 on Feb 5, 2011 7:59 AM

  • Table maintenence Generator for a Huge data Table.

    Hi Experts,
    I have created a Table maitenance for a Custom table which has some 80,000,000 records approx.
    Now the when i run this, it goes for a short dump saying "STORAGE_PARAMETERS_WRONG_SET".
    The basis have reported that the tcode is running a sequential read on the table & they are saying that is the reason for this dump.
    Are there any limitations that Table maintenance can't be created for table with huge data?
    Or should the program be modified to accomodate the "READ" from the tables in case of large entries?
    Please advice.
    regards,
    Kevin.

    Hi,
      I think this is because of memory overflow.
      You can create two screens for this, in on screen (Overview) screen, restrict the data selection.
      In detail screen display the data.
    With regards,
    Vamsi

  • Best way to show Reports

    Hi All !
    I am new to this Java developer forum, and trying to find the Best way to Display Reports from the Data base as cristal Reports,
    Is any easy way to implement this in my Java Project?
    with Regards
    vinoth

    Hi,
    This forum is exclusively for questions related to Sun Java Studio Creator. Could you pls post your question in an appropriate forum.
    Thanks,
    RK.

  • Best way to generate ER diagram

    We have Oracle 11g in our environment. I wanted to know which should be the best way to generate an ER diagram. Below are these steps in SQL developer (version of SQL developer is 3.2.x):
    a) View → Data Modeler → Browser
    b) In browser tab, select show on newly created relational model
    c) Then just drag the tables you want (from e.g. the «Connections» tab) onto the model.
    The problem is I am not able to drag the tables.
    I hope I have been able to explain my requirement of generating ER diagram.
    Please revert with the reply to my query.
    Regards

    2) If all the tables are required in ER diagram, then should all the tables be selected in point c) of my post.
    If all the Tables in your relational diagram are to included in the ER diagram, just click the Engineer button.
    If some are to be excluded, you can expand the Tables node in the tree in the top part of the dialogue and deselect the ones you don't want.  Then click the Engineer button.
    1) How to save the ER diagram
    To save your model for future update, you can go to the File menu and select Data Modeler > Save.
    To save your diagram as a PDF or image file, go to the File menu and select Data Modeler > Print Diagram.
    David

  • I have a PPC iMac 10.4.11 and will shortly buy a new i Mac.  What is the best way to transfer all my HD date from old to new? Thank you!

    I have a PPC iMac 10.4.11 and will shortly buy a new i Mac.  What is the best way to transfer all my HD date from old to new? Thank you!

    Migrating from PPC Macs to Intel Macs:
    https://discussions.apple.com/docs/DOC-2295
    How to use Migration Assistant:
    http://support.apple.com/kb/HT4413?viewlocale=en_US
    http://support.apple.com/kb/TS1963
    Troubleshooting Firewire target disk mode:
    http://support.apple.com/kb/HT1661

  • My Mum( a pensioner) is wanting to purchase both an iPhone and iPad.  What is the best way for her to manage her data /calls/txt etc. obviously the cost needs to be as low as possible.  Currently does not have WiFi but uses dongle

    My Mum( a pensioner) is wanting to purchase both an iPhone and iPad.  What is the best way for her to manage her data /calls/txt etc. obviously the cost needs to be as low as possible.  Currently does not have WiFi but uses dongle

    My Mum( a pensioner) is wanting to purchase both an iPhone and iPad.  What is the best way for her to manage her data /calls/txt etc. obviously the cost needs to be as low as possible.  Currently does not have WiFi but uses dongle

  • What is the best way to transfer all of my data from my iPod touch 3rd generation to my iPhone 4s?

    I have a 3rd generation iPod touch, and now I am getting an iPhone 4s. What is the best way to transfer all of my data, including paid-for apps and music, from the iPod to the iPhone without losing anything? Can I just do it through iTunes or the iCloud? I have never used iCloud before.

    Follow the directions here:
    http://support.apple.com/kb/HT2109

  • Best way to update individual rows of a Table?

    I've taken a look at some examples, though haven't gotten any clarification on this.  I am looking to have something close to a listbox or table to where I can update just a single column of row values at a 1 time per second pace.  I am looking to display our data-acquisition values in a table or listbox.  The single listbox seemed to work good for this, but I was unable to use row headers to list the channel names next to the channel values.  I was thinking about connecting the cursor values of two list-boxes to do this, but didn't find any info on this for the single list-box.
    I have a few questions:
    1) I have a 1D array to where I want to use that array of data to constantly update the first column (with a multitude of rows) of a table.  I am looking for the best route so as not to take up too much processing time in doing this.
    What is the best way to update individual rows of a table?   Invoke Node "Set Cell Value" ... or is there another method?
    2) Why is it that after every other iteration the row values are erased? 
    Also, for adding additional strings to the original arrray ... is it best to use the "Array Subset" and then the "Build Array" function, or the "Array Subset" and "Insert Into Array" function?
    See the attached example.
    Thanks.
    Solved!
    Go to Solution.
    Attachments:
    Table Example.vi ‏19 KB

    Jeff·Þ·Bohrer wrote:
    2) Why is it that after every other iteration the row values are erased?
    Classic race condition.  dump the for loop and p-node and just wire the 2D array to the table terminal.!
    I'm not seeing the race condition.  What I am seeing is the table emptying after the last element was written to it on every other run.  I saw watched this with highlight execution on.
    But I'm in full agreement with just writing to the terminal.  It is a 1D array, so you will need to use a build array and transpose 2D array in order for it to write properly.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Best way to change partition key on existing table

    Hi,
    Using Oracle 11.20.3 on AIX.
    We have a table about 800 million rows and 120gb in size.
    Want to try copies oif this table to evalaute different partitiong strategies.
    What is the quickest way to do this?
    Would have liked say datapump table 1 tro disk and datapumo import the data to new table but do the tables need to be of the same format.
    Thanks

    >
    Using Oracle 11.20.3 on AIX.
    We have a table about 800 million rows and 120gb in size.
    Want to try copies oif this table to evalaute different partitiong strategies.
    What is the quickest way to do this?
    Would have liked say datapump table 1 tro disk and datapumo import the data to new table but do the tables need to be of the same format.
    >
    First your subject asks a different question that the text you posted: Best way to change partition key on existing table. The answer to that question is YOU CAN'T. All data has to be moved to change the partition key since each partition/subpartition is in its own segment. You either create a new table or use DBMS_REDEFINITION to redefine the table online.
    Why do you want to export all data to a file first? That just adds to the time and cost of doing the op.
    What problem are you trying to use partitioning to solve? Performance? Data maintenance? For performance the appropriate partitioning key and whether to use subpartitions depends on the types of queries and the query predicates you typically use as well as the columns that may be suitable for partition keys.
    For maintenance a common method is to partition on a date by year/month/day so you can more easily load new daily/weekly/monthly data into its own partition or drop old data that no longer needs to be kept online.
    You should use a small subset of the data when testing your partitionings strategies.
    Can you do the partitioning offline in an outage window? If not then using the DBMS_REDEFINITION is your only option.
    Without knowing what you are trying to accomplish only general advice can be given. You even mentioned that you might want to use a different set of columns than the curren table has.
    A standard heap table uses ONE segment for its data (ignoring possible LOB segments). A partitioned/subpartitioned table uses ONE segment for each partition/subpartition. This means that ALL data must be moved to partition the table (unless you are only creating one partition).
    This means that every partitioning scheme that uses a different partition key requires ALL data to be moved again for that test.
    Provide some information about what problem you are trying to solve.
    >
    Is this quicker than datapump?
    >
    Yes - exporting the data simplying moves it all an additional time. Ok to export if you need a backup before you start.
    >
    Found artcle which talks about using merge option on datapump import to convert partitioned table to non-partitioned table.
    >
    How would that apply to you? That isn't what you said you wanted to do.

  • Best way to generate one record per day from a table with eff/exp dates

    Hi,
    Have a table which has various attributes and an eff and exp date. e.g attributea, 01/05/2012, 16/05/2012
    We wish to create another table from this table to have one record per day. e.g 16 records.
    What is best way to achieve this in OWB ?
    Thanks

    Hi,
    Example if have table
    with following contents
    conversion_rate number(6,4)
    EFFEcTIVE_DATE DATE
    expiration_date date
    example record 1.43, 01/05/2012,16/05/2012
    If want to have another table which instead has 16 records one for each day
    e.g
    1.43, 01/05/2012
    1.43,02/05/2012
    1.43,16/05/2012
    Thoughts on best way to do this.
    Thanks

  • The best way to get reports 6i output into Excel 97

    Hi there,
    I have a number of graphical reports that have a few group levels. These generate HTML or PDF output ok.
    I want to get the output to look as close to the HTML/PDF output from Report Server or Runtime into Excel 97.
    Outputting as HTML and importing into Excel does not work very well as Excel places all the horizontal data into one cell! This works perfectly in Excel 2000 but our client does not have this licenced so it is not a solution.
    I have thought about outputting as delimited output. This has the effect of placing all the parent group information on each detail line (it is like the group by clause has been removed if you did something similar in SQL*Plus) along with all the prompts and it looks a real mess and takes up many columns in Excel 97.
    What is the best way to approach this?
    Cheers,
    John

    If this is a client/server app, then you could use ole or dde to put the data directly into Excel.
    Chad

  • What is the best way to keep Drivers up to date?

    Hi Guys
    I am looking to keep a range of makes and models drivers up to date. I have done a bit of reading through a few blogs and wondering if DPInst.exe through a required task sequence is the best way?; if SCUP 2011 is useable?;
    or is there a better way?
    Thanking you in advance
    NN

    I have done a post on the subject myself. The blog is mainly for my own purpose, but I hope you get the basics of it:
    http://www.imab.dk/extending-hardware-inventory-in-sccm2012/
    Martin Bengtsson | www.imab.dk
    Thats very interesting, thanks you.
    NN
    Oh, you might need to know how to use regkeytomof, which is the tool that generates the .mof edits you need:
    http://be.enhansoft.com/post/2014/02/13/How-to-Use-RegKeyToMof.aspx
    Martin Bengtsson | www.imab.dk

  • Best way to store, present en input date/time

    Hi!
    I hope I can make myself clear. I'm designing an APEX application that has a couple of different date-type fields. That application is going to be translated in two different languages, that has different date/time representations. For example: dd-mon-yyyy hh12:miAM and dd-mm-yyyy hh24:mi. I'm now finding out what's the best way to store those date/time fields and hou I can present them in a report/form and how the user can enter them in a field.
    This is what I see:
    - I can set an application date format/application timestamp format. This has to be the in the format of the main language, in this case, English.
    - I also can make translated versions of the application. QUESTION: can I enter a different date format for the translated application?
    Do I have to make my fields of the date type or timestamp type?
    I hope someone can answer. If I'm not clear, please tell me!

    You can set the application date format using an application item
    &DATE_FORMAT_MASK.
    and set this item dynamically.
    Home>Application Builder>Application ####>Shared Components>Edit Globalization Attributes
    Denes Kubicek
    http://deneskubicek.blogspot.com/
    http://www.opal-consulting.de/training
    http://apex.oracle.com/pls/otn/f?p=31517:1
    http://www.amazon.de/Oracle-APEX-XE-Praxis/dp/3826655494
    -------------------------------------------------------------------

  • What is the best way to read and manipulate large data in excel files and show them in Sharepoint

    Hi ,
    I have a large excel file that has 700,000 records in it. The excel file has a few columns that change every day.
    What is the best way to read the data form the excel file in fastest and most efficient way.
    2 nd Problem,
    I have one excel file that has many rows each row contain some data that has certain keywords.
    What I want is  to segregate the data of rows into respective sheets(tabs ) in the workbook.
    for example in rows have following data 
    1. Alfa
    2beta
    3 gama
    4beta
    5gama
    6gama
    7alfa
    in excel
    I want there to be 3 tabs now with each of the key words alfa beta and gamma.

    Hi,
    I don't really see any better options for SharePoint. SharePoint use other production called 'Office Web App' to allow users to view/edit Microsoft Office documents (word, excel etc.). But the web version of excel doesn't support that much records as well
    as there's size limitations (probably the default max size is 10MB).
    Regarding second problem, I think you need some custom solutions (like a SharePoint timer job/webpart ) to read and present data.
    However, if you can reduce the excel file records to something near 16k (which is supported rows in web version of excel) then you can use SharePoint Excel service to refresh data automatically in the excel file in SharePoint from some external sources.
    Thanks,
    Sohel Rana
    http://ranaictiu-technicalblog.blogspot.com

Maybe you are looking for