Table with a huge size. With big fields... should be in a separated table?

Hi.
Consider the following table
id_foo num_doc_foo pages_foo pdf_foo .. etc
imagine that the table has millions of records, and a total size of 4GB or more.
im collecting data for build a statistical form, but the query takes forever, althow analizing the query all checks out.
My question is, the size of the table might influence the query performance?
Should i create a separated table to collect the values i want, and keep the previous table to save the data (the pdf_data documents are huge).
Resuming what im saying.
A table for the huge data fields (with a big size), and another table that connects to the previous one, but with smaller data, data for generating reports, etc.
What do u think? Does it makes sence?
I work with oracle 9.2x , and can't upgrade.
Cheers.

don't remember exact details for 9.2, but I believe that oracle stores the LOB separate from the rest of the table, if it's over a certain threshold. so, if the pdfs are large enough, they aren't really stored with the table data, so reporting queries won't be affected. check the manual for how LOB storage works, and then store the pdfs in a separate table if needed.

Similar Messages

  • Messages are not dequeuing from ECX_outqueue table resulting in huge size of database.

    EBS Tech Stack :- EBS R12
    Release :- 12.1.3
    Oracle Database 11.2.0.2
    SQL> select count(*) from ECX_OUTQUEUE;
    COUNT(*)
    118405
    We are not using OTA(Oracle Transport Agent) in our current setup. This problem started from September 3, 2013. Also we are not using any XML gateway for this.
    Kindly let us know How I can get that what is the origin of the data and why its not dequeing.

    Are there any exceptions in the table? -- http://etrm.oracle.com/pls/et1211d9/etrm_pnav.show_object?c_name=ECX_OUTQUEUE&c_owner=APPLSYS&c_type=TABLE
    Have you tried to rebuild the related queues as per https://forums.oracle.com/thread/2593111 ?
    If this doesn't help, please log a SR.
    Thanks,
    Hussein

  • Use of block sizes with APO pgm /SAPAPO/RDMCPPROCESS

    I am using APO V5.1.
    I am using standard pgm /SAPAPP/RDMCPPROCESS to send APO planning data back to the connected ECC system.
    One of the parameters is block size.
    I am using the default of 1000.
    But what are the performance implications of using different block sizes? What is 'good practice' for using this block size parameter?
    Thanks for any advice on this point...

    Hi,
    The F1 help itself gives good information on this topic.
    The block size indicates how many change pointers are processed in one shot (one block).
    So, if you increase the block size, it means that more change pointers are processed in one shot meaning that the memory requirement would be higher. On the other hand, since you have increased the change pointers to be processed in one shot, the number of blocks required would be less. So, if your system has sufficient capacity to process the number of change pointers that you specify in a block, increasing the block size would increase the load on the memory, but would make the process fast, meaning that your change pointer job would finish faster.
    Vice versa is also true.
    Your system performance (mostly linked to memory availability) could get negatively impacted if you increase the block size. If you see that your system performance goes down with this block size of 1000, this should mostly mean that your memory resources are already exhausted in other processes, and you could reduce the block size to 500. With this, you free up memory and system would perform faster. Though your change pointer job would take a little bit more time.
    Try first with having the default block size. Unless you face some issue, there's no point in changing the default value.
    Hope this explains.
    PS: In our case, we mostly use 1000, but during some peak system load times, we use 500 as block size.
    Thanks - Pawan

  • APQD table size is fastly increasing with huge size

    Hi All,
    File system space has been increasing very fast. We analyzed
    the tables for the large size, from which we found APQD table size is
    relatively very large in size in Production system. So we are planning
    to delete old session logs till last 5 days from the APQD table. We
    reffered the note
    number "36781 - Table APQD is very large", and run the
    report RSBDCCKA and RSBDCCKT and then "RSBDCREO", but found no
    significant change at table level and file system as well. (attached is
    the screenshot for the same).
    We are planning to re-organize the APQD table, so that we can have free
    size. Can you please help me to know, reorganization of APQD table will
    help to gain segnificant space? We have APQD table size = 350 GB. So
    the re-org will generate huge archive log files which needs to be taken
    backup. To avoid this can we have other strategy to tackle this problem?
    Quick help is appriciated to sort out this problem in production
    environment.
    -Anthony...

    > number "36781 - Table APQD is very large", and run the
    > report RSBDCCKA and RSBDCCKT and then "RSBDCREO", but found no
    > significant change at table level and file system as well. (attached is
    As far as I remember, RSBDCREO is client dependent.
    If you did run the report in client 000 it might not affect the old
    batchinput data in your business client.
    A reorg only makes sense, if you did a significant amount of deletes before.
    As for filesystem space, you should see a reduced amount of files in
    the batch input subdirectories in GLOBAL_DIR.
    Volker

  • Hi, I have big problems since 2 weeks with outline, fonts, size in Ai CS6 : Numerical value are so big

    Hi, I have big problems since 2 weeks with outline, fonts, size in Ai CS6 : Numerical value are so big and I can't change it with smaller. Apparently Ai said me that comes from AiPrefs... Could you help me please.
    I tried to initialize my preferences but nothing change.
    PS : Sorry for my English, I'm French
    Pierre

    Pierre,
    You may try the list and see how far down you need to go; you have probably been through 1) and 2), and 4) is both reversible and more thorough than 3); 5) is hardly relevant, so you may end up with 6).
    The following is a general list of things you may try when the issue is not in a specific file (you may have tried/done some of them already); 1) and 2) are the easy ones for temporary strangenesses, and 3) and 4) are specifically aimed at possibly corrupt preferences); 5) is a list in itself, and 6) is the last resort.
    If possible/applicable, you should save current artwork first, of course.
    1) Close down Illy and open again;
    2) Restart the computer (you may do that up to at least 5 times);
    3) Close down Illy and press Ctrl+Alt+Shift/Cmd+Option+Shift during startup (easy but irreversible);
    4) Move the folder (follow the link with that name) with Illy closed (more tedious but also more thorough and reversible);
    5) Look through and try out the relevant among the Other options (follow the link with that name, Item 7) is a list of usual suspects among other applications that may disturb and confuse Illy, Item 15) applies to CC, CS6, and maybe CS5);
    Even more seriously, you may:
    6) Uninstall, run the Cleaner Tool (if you have CS3/CS4/CS5/CS6/CC), and reinstall.
    http://www.adobe.com/support/contact/cscleanertool.html

  • Reg - Reading internal table with multiple record in a single field

    Dear Guru's,
                        i want to read a internal table with field having mutilple entries like
    read table READ TABLE LT_T2 INTO LT_T2_WA WITH KEY HKONT IN ( '0001152430', '0001152930', '0001152410' ).
    But it says comma without preceding colon (after READ?).
    please guide me.....
    thanks & Regards,
    Balaji.S

    ya this is inside the loop.
    plz check....
    loop at lt_t2 into lt_t2_wa.
    READ TABLE LT_T2 INTO LT_T2_WA WITH KEY HKONT IN ( '0001152430', '0001152930', '0001152410' ).
    endloop.
    thanks & Regards,
    Balaji.S

  • Table for Analysis authorization along with values for authorization fields

    Hi,
    I am looking for table that contains the Analysis Authorization name along with values for all the authorization fields within this Analysis Authorization. Individually i can go to PFCG or Rsecadmin but since i need all the Analysis auth objects, i need to get this info into excel, so need a table.

    Hi Prashanth
      You can check RSECVAL that is appropriate for your requirement please let us know if any further help is needed.
    Thanks & Regards
    Santosh Varada

  • Hi guys n girls. How do you copy a whole table to create a new table with all cell sizes in tact? Thanks for your help. Jason.

    Hi guys n girls. How do you copy a whole table to create a new table with all cell sizes in tact? Thanks for your help. Jason.
    when you copy n paste into a new table, all the cell sizes are changed.
    is there a way to put in a new table from your templates into an existing file, different to the standard very basic ones in insert table.
    I look forward to your answers.  Your help is very much appreciated.
    Also how do you search for question answers already written in this support area please.

    Hi Jason,
    In Numbers 3, you can select a whole table by clicking once in the table to make it active, then click once on the "bull's eye" at the top left.
    Now copy and paste. All formatting (and any cell content) is pasted intact. In Numbers 2.3 (Numbers '09) it is a little different for selecting a whole table. But I won't go into that unless you are using Numbers '09. Please reply.
    I don't like the look of the tables in Insert Table. I keep custom tables in My Templates. I have set Numbers > Preferences > General > For New Documents > Use template: (name of my favourite custom template)
    That opens when I launch Numbers, or ask for a new document (command n). Note that if you follow this preference setting, then Menu > File > New From Template Chooser (for another template) requires you to hold down the option key in that menu.
    Regards,
    Ian.
    Message was edited by: Yellowbox. All formatting (and any cell content) is pasted intact.

  • Right way to preserve all parent table entries in a join with many tables!!

    This problem is quite interesting to me. I have asked this question to others but no body is able to provide me with proper answers.
    The problem is: How do I join a huge parent table with many child tables (more than 5 child tables) preserving all of the parent table entries. Lets say there is the parent table parentTable and three child tables childTable1, childTable2, childTable3. In order to get the data after joining these tables the query that I have been using was:
    select parent.field1, parent.field2, parent.field3, child1.field4, child1.field5, child2.field6, child3.field7 from ParentTable parent, childTable1 child1, childTable1 child2, childTable3 child3 where parent.fielda = child1.fieldb and parent.fieldc = child.fieldd and parent.fielde = child.fieldf.
    Although the tables are huge (more than 100,000 entries), this query is very fast, however those parent table entries which do not have child entries are lost. I know that I can left join a parent table with a child table and then with the next child table and then with the next child table and continue. Isn't there a simple solution for this commonly happening problem?
    Please provide suggestions please...

    Hello Lakshmi,
    Although I do not know exactly how to achieve what you want to achieve but I have seen DBAs/ABAPers in my experience can run queries/scripts using the COUNT function to give the actual number of line items per table for all your 100 or odd tables.
    Rgds
    CONMJI

  • How to know the specific size of a field in oracle table

    Hello,
    oracle 11.2.0.3 on redhat
    i have three tables with same structure, table0, table50, table100:
    create table table0 (id number, shipment RAW(2000), options RAW(2000));
    i inserted 10000 rows in each but with some differences:
    table0 has 10000 id, but shipment and options has null values
    table50 has 10000 id, but shipment and options has only 5000 rows, the other 5000 are null values
    table100 has 10000 rows of id, shipment and options. No null values.
    I want to know the specific size of each column, also the size of each row, to know the exactly size of the field.
    for example:
    row 1, column options of table100 has "asdfasdfagasdgasbabsdgoasdpgiahnwe1129u412094u12"
    row 2, column options of table100 has "a".
    both are raw datatype, but i think its space will has different size (bytes), how to know it? any query?
    I found this query:
    select owner,tablespace_name,segment_name,sum((bytes/1024/1024)) Bytes
    from sys.dba_extents
    where owner in 'MAA' and segment_type='TABLE' and segment_name ='TABLE00'
    group by tablespace_name,owner,segment_name
    order by owner,tablespace_name,segment_name, bytes; but i think it isn't my solution.

    thx mr. sb92075
    i tried with vsize too but same result:
    with table50
    SELECT VSIZE(OPTIONS) FROM TABLE00;
    VSIZE(OPTIONS)
                 1000
                 1000
                 1000
                 1000
                   *with table100
    SELECT VSIZE(OPTIONS) FROM TABLE100;
    VSIZE(C_ELECCION)
                 2000
                 2000
                 2000
                 2000
                      *it's the same size for 'a' like 'abcasdfjañldngaosnbaposnbpaosuqroiwejrklasndiaoshbaposdjgajs'?? It's hard for me to believe.
    data is saved in blocks, a extents is a contiguos number of blocks and a group of extensions are segments. If you insert some text in a table field, Oracle create a extension and insert there the data but depend the bytes you insert Oracle will create one or more extents. How to know this specific space in that field?

  • Create materialized view with specific column sizes

    Hi all,
    I'm trying to create a materialized view with a specific a column size. Something like
    create materialized view test_mv
    refresh force on demand
    as
    select id,
           cast(my_compound_field as nvarchar2(50))
    from ( select id,
                  field1 || field2 my_compound_field
           from   my_table);But Oracle seems to ignore the cast and takes the maximum size it finds for field1 || field2 in the select query. The resulting table has a column nvarchar2(44) instead of nvarchar2(50).
    This can give a problem when the view is refreshed... there could be new data that exceeds the current size, i.e. where length(field1 || field2) > 44.
    How can I override the column size of a field in a materialized view?
    Edit: Some additional info to clarify my case:
    field1 and field2 are defined as nvarchar2(25). field1 || field2 can theoretically have a length of 50, but there is currently no data in my table that results in that length, the max is 44. I am afraid that there will be data in the future that exceeds 44, resulting in an error when the MV is refreshed!
    Edited by: Pleiadian on Jan 25, 2011 2:06 PM

    Cannot reproduce what you are saying is happening.
    SQL> create table t (a nvarchar2(50), b nvarchar2(50));
    Table created.
    SQL> create materialized view tmv as
      2  select a, b, a || b c from t;
    Materialized view created.
    SQL> desc tmv
    Name                                      Null?    Type
    A                                                  NVARCHAR2(50)
    B                                                  NVARCHAR2(50)
    C                                                  NVARCHAR2(100)
    SQL> drop materialized view tmv;
    Materialized view dropped.
    SQL> create materialized view tmv as
      2  select a, b, substr(a || b, 1, 10) c from t;
    Materialized view created.
    SQL> desc tmv
    Name                                      Null?    Type
    A                                                  NVARCHAR2(50)
    B                                                  NVARCHAR2(50)
    C                                                  NVARCHAR2(10)
    SQL> select * from v$version;
    BANNER
    Oracle Database 11g Enterprise Edition Release 11.1.0.7.0 - 64bit Production
    PL/SQL Release 11.1.0.7.0 - Production
    CORE    11.1.0.7.0      Production
    TNS for Linux: Version 11.1.0.7.0 - Production
    NLSRTL Version 11.1.0.7.0 - Production
    SQL>Edited by: 3360 on Jan 25, 2011 8:10 AM
    And with data
    SQL> insert into t values ('3123423423143hhshgvcdcvw', 'ydgeew  gdfwe   dfefde  wfjjjjjjj');
    1 row created.
    SQL> commit;
    Commit complete.
    SQL> desc tmv
    Name                                      Null?    Type
    A                                                  NVARCHAR2(50)
    B                                                  NVARCHAR2(50)
    C                                                  NVARCHAR2(10)
    SQL> select * from tmv;
    A
    B                                                  C
    3123423423143hhshgvcdcvw
    ydgeew  gdfwe   dfefde  wfjjjjjjj                      3123423423

  • Change the Size of  "Code"  field in Master Table of UDO

    Hi there,
    I am using UDO in my form, where "Code" is the Primary key field of the Master Table.
    Now the Size of the same is 8 characters. Can any one of you help me, to change the "Code" size to 15 characters?
    Thanking you,
    Pradeep

    Hi pradeep,
    You can't. But you can create a new alphanumeric field with the intended size.
    Regards,
    Vítor Vieira

  • Best way of partitioning the huge size table in oracle 11gr2

    hi,
    OS: linux
    DB: oracle 11gR2
    table size is about 2T and single table space with multiple dbf files in ASM,
    table partition with dbms.redefinition is running past 2 days (range partition).. its running but not that fast..
    note: exchange partition is too slow.. hence not used..
    what is the best way of doing the huge size table partition. please suggest, thanks.

    >
    what is the best way of doing the huge size table partition
    >
    A few questions
    1. Is that an OLTP or OLAP table?
    2. Is all of the data still needed online or is some of it archivable?
    3. What type of partitiioning are you doing? RANGE? LIST? COMPOSITE?
    4. Why do you you say 'exchange partition is too slow' - did you do a test?
    For example if data will be partitioned by create date then one stragety is to initially put all existing data into the root partition of a range partitioned table. Then you can start using new partitions for new incoming data and take your time to partition the current data.

  • Impact of Changing Data Package Size with DTP

    Hi All,
    We have delta dtp to load data from DSO to infocube. Default data package size with dtp is 50,000 records.
    Due to huge no of data, internal table memory space is used and data loading get fails.
    Then we changed the data package size to 10,000, which executes the data load successfully.
    DTP with package size of 50,000 took 40 minutes to execute and failed, but DTP with package size of 10,000 took 15 minutes (for same amount of data).
    Please find below my questions:
    Why a DTP with bigger size of packet runs longer than a DTP with lower packet size ?
    Also by reducing the standard data package size 50,000 to 10,000, will it impact any other data loading?
    Thanks

    Hi Sri,
    If your DTP is taking more time then check your transformation .
    1.Transformation with Routines always take more time so you if you want to reduce the time of execution then routine should be optimized for good performance .
    2.Also check if you have filter at DTP level .Due to filters DTP takes long time .If same data get filtered at routine level it take much lesser time .
    3.If you cannot change routine then you can set semantic keys at your DTP .The package data will be sorted as per semantic keys and thus it may be helpful at routine level for fast processing.
    4.Your routine is getting failed due to  internal table memory space so check if you have select statement in routine without FOR ALL ENTRIES IN RESULT_PACKAGE or SOURCE_PACKAGE line .if you will use this It will reduce record count .
    5.Wherever possible delete duplicate records and if possible filter useless data at start routine itself .
    6.Refresh internal table if data no longer needed .If your tables are global then data will be present at every routine level so refreshing will help to reduce size.
    7.The maximum memory that can be occupied by an internal table (including its internal administration) is 2 gigabytes. A more realistic figure is up to 500 megabytes.
    8.Also check no of jobs running that time .May be you have lots of jobs active at the same time so memory availability will be less and DTP may get failed .
    Why a DTP with bigger size of packet runs longer than a DTP with lower packet size ?
    *Start and end routine works at package level so routine run for each package one by one .By default package have sorted data based on keys (non unique keys (characteristics )of source or target) and by setting semantic keys you can change this order.So Package having more data will take more time in processing then package have lesser data .
    by reducing the standard data package size 50,000 to 10,000, will it impact any other data loading?
    It will only impact running of that load .but yes if lots of other loads are running simultaneously then server can allocate more space to them .So better before reducing package size just check whether it is helpful in routine performance (start and end ) or increasing overhead .
    Hope these points will be helpful .
    Regards,
    Jaya Tiwari

  • Problem with Refreshing the data bound text fields

    Folks,
    First of all, this a GREAT product and I am moving my apps from Eclipse to Studio Creator. I am very impessed with it so far.
    I am gone through most tutorials, but can't help resolve the problem that I am having. I have a page, that lists all the users in a table. When you click on any user, it opens up another page that lists the detail of the specific user you have clicked on. This a typical use in most web applications, right?
    In the user detail page, I open it in "Read" only mode, by setting the "setReadOnly" to true for all text field components. Then users can edit it by clicking "Edit" button. Edit button just changes the "setReadOnly" to false, to allow editing. After editing, they can hit the "Save" button to save the changes. In the save button, I call this code:
    public void prerender() {
            try {
                userID=getSessionBean1().getUserID();
                getSessionBean1().getWeb_userRowSetUserDetail().setObject(1,userID);
                getWeb_userDataProviderUserDetail().refresh();
                info("prerender: Refresh called");
            } catch (Exception ex) {
                error("Error in userForm.prerender():"+ex.getMessage()+ex.getStackTrace());
                log("Error in userForm.prerender():", ex);
            setReadWrite();
    public String btnSave_action() {
            try {
                getWeb_userDataProviderUserDetail().commitChanges();
                getWeb_userDataProviderUserDetail().refresh();
                info("Data Provider Refreshed in SAVE");
                _readOnly=true;
                form1.discardSubmittedValues("saveVForm");
                info("User Updated New:"+userID);
            } catch (Exception ex) {
                log("Error Description", ex);
            return null;
        public String btnEdit_action() {
            try {
                _readOnly=false;
                //info("User ID Edt="+getSessionBean1().getUserID());
            } catch (Exception ex) {
                error("Error in userForm.btnEdit_action:"+ex.getMessage()+ex.getStackTrace());
                log("Error in userForm.btnEdit_action():", ex);
            return null;
        private void setReadWrite(){
            // Set read-only to true or false for all text fiels
            this.btnEdit.setVisible(_readOnly);
            this.btnSave.setVisible(!_readOnly);
            this.user_id.setReadOnly(_readOnly);
            this.user_email_client.setReadOnly(_readOnly);
            this.user_email_office.setReadOnly(_readOnly);
            this.user_fname.setReadOnly(_readOnly);
            this.user_lname.setReadOnly(_readOnly);
            this.user_id.setReadOnly(_readOnly);
            this.user_password.setReadOnly(_readOnly);
            this.user_password_conf.setReadOnly(_readOnly);
            this.user_phone_cell.setReadOnly(_readOnly);
            this.user_phone_cell_aac.setReadOnly(_readOnly);
            this.user_phone_client.setReadOnly(_readOnly);
            this.user_phone_home.setReadOnly(_readOnly);
            this.user_phone_office.setReadOnly(_readOnly);
        }When the save button is clicked, the database is updated (as I can verify by looking directly into the database), but the data-bound text fields do not display the new value. YES, I am calling refresh() on data provider? I tried to call refresh in both "preRender()" and "Save" button action event, but that did not help either.
    Please note that the save button also changes the "setReadOnly" back to TRUE for all text fields. That means, users view the page in Read-Only mode after saving. (They have to click on "Edit" button again to edit it). I discovered that if I do not set the text field components in "ReadOnly" mode, then they display the updated value of the fields. But if I change setReadOnly to TRUE, then they display the old values (before the save).
    Any idea what I am doing wrong here?
    Thanks

    Here is my experience with it. I have three text
    boxes to be filled out. They are part of a virtual
    form with an add button as the submit. You can fill
    out the 3 and add them and a node is added to a tree
    component. When you click on a node in the tree
    component (the submit to another virtual form that
    the 3 buttons participate in) the text fields should
    be loaded with the values that they where added with.Hard to tell from this explanation, but if the text fields do not participate in a virtual form, then they are not going to get updated when an action in the virtual form happens. Not sure this is your problem. Like I say, hard to tell from this explanation.
    They will not display the values though...having
    nothing to do with read only. If I disable them
    before hand...they display the updated values. If I
    set their value at the top or bottom of prerender,
    they display the change...if I set the value in an if
    conditional that is hit (I have checked many times)
    in prerender, they will not display the values. If I
    set them in the tree handler anywhere, they will not
    display the values. This is very frustrating and is
    wasting tons of my time...I simply want to load the
    text fields based on the tree node that was
    clicked...I have messed with this for hours and it
    certainly does not work.Well, setting them in the tree handler won't work because the tree handler gets called in the page instance that handles the submit and not the page instance that renders the response.
    Posting your prerender code might help. Also, your action handler.
    >
    >
    - Mark

Maybe you are looking for

  • Why am i getting an error message saying i don't have enough privileges to access

    When i try and sync my ipod to my itunes it give me that error message. And now there are multiple songs with "!" on them. Why is this?

  • Regarding Message Maaping

    HI, I have a scenarion where i have to read a file where the structure is not defined properly i.e Records are coming as continous stream of messages separated by some special ASCII charac but in that stream there is only small portion of useful info

  • Finding Errors in /var/adm/messages file

    Hi, I am new to UNIX admin, i am going to write a script in such a way that it has to send a mail to root if any errors in /var/adm/messages file. Can any one please send useful links or sample script file? Thanks Ramesh

  • Canon 350D/450D vs. Photoshop CS4/5

    I wonder how to setup a Canon camera with USB to a laptop and get the picture auto into photoshop/Bridge when the picture is taken. Anyone who have done this?

  • Please Help my skype might be broken

    My Skype is looking really weird look at what it looks like  help me figure out how to fix it please