Database size: optimization: tuning

Hi:
I'm using Berkeley DB as it is supplied with MacOS 10.5.6
on an intel iMac with 2G of ram. On this machine, as well as on
a more powerful machine with more memory I see the following:
I'm loading a tied hash via Perl BerkeleyDB.pm with constant-length
keys, and values that vary from 20 to 200 bytes.
Loading time is constant at about 0.9 seconds per 25,000 records
(0.6 per 25,000 on the more powerful machine)
until the dbm file reaches about 660 MB (1.2 million records).
Then there is an inflection point and loading time increases
so that at 2 million records (dbm file =710 MB) it is 33 seconds per
25,000 records. If i reduce the average size of the values of each record, the inflection
point stays constant at ~660 MB, although more records are stored.
I wonder if there are database tuning/optimization functions I should
be looking at.
Thanks
Richard Moe

There is in fact no better way than doing the real test yourself, but you can also just do the math. Start with figuring/estimating what's going to be the load of the application, how many queries are to be executed during one request/session, etcetera.

Similar Messages

  • Use OEM to monitor oracle database size

    Hello All,
    We currently have several database which are currently monitored via OEM however i will appreciate if anyone can tell me how i can monitor the the size of my oracle databases via oem
    Thanks

    Hi,
    I'm not sure of OEM will monitor database size, but many 3rd party tools do.
    There are many ways to achieve Oracle table monitoring, and many create specialized extension metadata tables to monitor Oracle table growth. In Oracle 10g you can perform Oracle tables monitoring with the dba_hist_seg_stat tables, specifically the space_used_total column.
    On my databases, I create STATSPACK extension tables to track table and overall database growth:
    http://www.dba-oracle.com/te_table_monitoring.htm
    The full scripts are in my Oracle Press book, Oracle tuning with STATSPACK . . .
    http://www.amazon.com/Oracle9i-High-Performance-Tuning-STATSPACK-Burleson/dp/007222360X
    Hope this helps . . .
    Donald K. Burleson
    Oracle Press author
    Author of "Oracle Tuning: The Definitive Reference"
    http://www.rampant-books.com/book_2005_1_awr_proactive_tuning.htm

  • Transaction Sync and Database Size

    Hello,
    We're using BDB (via the Java interface) as the persistent store for a messaging platform. In order to achieve high performance, the transactional operations are configured to not sync, i.e., TransactionConfig.setSync(false) . While we do achieve better performance, the size of the database does seem rather large. We checkpoint on a periodic basis, and each time we checkpoint, the size of the database grows, even though records (messages in our world) are being deleted. So, if I were to add, say 10000 records, delete all of them and then checkpoint, the size of the database would actually grow! In addition, the database file, while being large, is also very sparse - a 30GB file when compressed reduces in size to 0.5 GB.
    We notice that if we configure our transactional operations to sync, the size is much smaller, and stays constant, i.e., if I were to insert and subsequently delete 10000 records into a database whose file is X MB, the size of the database file after the operations would be roughly X MB.
    I understand that transaction logs are applied to the database when we checkpoint, but should I be configuring the behaviour of the checkpointing (via CheckpoinConfig )?
    Also, I am checkpointing periodically from a separate thread. Does BDB itself spawn any threads for checkpointing?
    Our environment is as follows:
    RedHat EL 2.6.9-34.ELsmp
    Java 1.5
    BDB 4.5.20
    Thanks much in advance,
    Prashanth

    Hi Prashanth,
    If your delete load is high, your workload should benefit from setting the DB_REVSPLITOFF flag, which keeps around the structure of the btree regardless of records being deleted. The result should be less splits and merges, and is therefore better concurrency.
    Here you can find some documentation that should help you:
    Access method tuning: http://www.oracle.com/technology/documentation/berkeley-db/db/ref/am_misc/tune.html
    Transaction tuning: http://www.oracle.com/technology/documentation/berkeley-db/db/ref/transapp/tune.html
    If you are filling the cache with dirty pages, you can indeed call checkpoint() periodically in the application, or you can create a memp_trickle thread. See the following sections of the documentation:
    - Javadoc: Environment -> trickleCacheWrite" http://www.oracle.com/technology/documentation/berkeley-db/db/java/com/sleepycat/db/Environment.html#trickleCacheWrite(int)
    Some related thread for the "database size issue", can be found here: http://forums.oracle.com/forums/thread.jspa?threadID=534371&tstart=0
    Bogdan Coman

  • SQL azure database size not dropping down after deleting all table

    Dear all,
    I have a simple database on Azure for which I have deleted all table data. The size of the database is still showing 5Mb of data and I am charge for that. I have heard that this may happen from cluster index getting fragmented.
    I have run  a querry I found on internet on all my table index to show percentage of fragmentation and all report 0%.
    DBA is not so my job but what could it be or how can I reduce that size ?
    ON premise I would use COMPACT DB but not available in azure like some others DB action
    Thnaks for tips
    regards

    user created objects/data are not the only ones stored in your database. you got system objects and metadata as Mike mentions above.
    are you trying to skip being charged if you're not storing data? looking at the pricing table,  you'll still get charged the $4.995 for the 0-100MB database size range.

  • Tables are deleted but database size does not change in sql server 2008r2

    Hi All,
    20GB Tables are deleted in my database but database size does not change and disk size showing same size.

    Hi ,
    I have ran the Disk usage by Top Tables report and Identified couple of tables with unwanted data for last 5 years. I have deleted the data for the first 3 years and then ran the Disk usage by Top Tables report again. When I compared the report before
    and after the data deletion, I have noticed certain facts which is not matching with what I know or learned from experts like you. The following are the points where I am looking for clarification:
    1.Reserved (KB) has been reduced. I was expecting the data Reserved (KB) will remain the same after the data deletion. 
    2. The Data(KB) and Indexes(KB) fields have been reduced as expected. The Unused(KB) field have been increased as expected.
    I was expecting the total of Data(KB) and Indexes(KB) field space gained will be equal to the Unused(KB) field gained after deleting the data. But that is not the case. When I deducted(subtracted) the difference in  the Reserved(KB)(Difference before
    and after data deletion) field from the Total of space gained from the data deletion is equal to the Unused(KB) gained field value.
    I am not a SQL expert and not questioning but trying to understand whether we really gain space by deleting data from the tables. Also keen to get the concepts right, but my testing by deleting some records confused me.
    Looking ahead to all your expert advice.
    Thanks,
    Vennayat

  • What is the best practice on mailbox database size in exchange 2013

    Hi, 
    does anybody have any links to good sites that gives some pros/cons when it comes to the mailbox database sizes in exchange 2013? I've tried to google it - but hasn't found any good answers. I would like to know if I really need more than 5 mailbox databases
    or not on my exchange environment. 

    Hi
       As far as I know, 2TB is recommended maximum database size for Exchange 2013 databases.
       If you have any feedback on our support, please click
    here
    Terence Yu
    TechNet Community Support

  • How to get the database size of several system

    Hi
    We've got a lot of database system on our landscape. I want to make a simple report to get the size the all our database on a weekly base.
    I try to get a RZ20 value on it without succes.
    Do you know a sql request that i could launch on all my database to get this information ? i did'nt find any V$^view with this information.
    thanks for your help
    florent

    "The database size" can be
    - the total size
    - the filled size
    - the filled size without undo/temp
    You can use DBACOCKPIT to centrally manage multiple instance and check the sizes.
    Markus

  • How to determine the database size corresponding to the nber records in DSO

    Hi Colleagues,
    I would like to determine the database size corresponding to my new BI project.
    I know the number of records uploaded in the DSO from the source system for the intialization phase.
    How can I deduct the database size / disk size corresponding to the number of record uploaded ?
    Thanks,

    Hi Ram,
    I am with SAP BI Release SAPKW70019
    I do not have the option Single Table analysis -
    I have in DB02 or ST04 the following options.
    *- Space*
    -- space overview
    ->database
    --overview
    ->users
    --overview
    --detailed analysis
    -> tablespaces
    --overview
    --detailed analysis
    -> segments
    --overview
    --Detailed analysis
    --Detailed Analysis Aggregated
    -> Additional Functions
    --Collector Logs
    --BW Analysis
    Where should I go through ?
    Thanks

  • Sharepoint 2010 - configuration database size limit

    Hi,
       We have below scenario in our production farm. Ours is a SharePoint 2010 Enterprise Edition server.
    Medium size farm (4 WFE, 4 App Servers, Content DBs in separate tier. No SharePoint search in this farm)
    Configuration DB reached 250+ GB. Please clarify below queries.
    1. What is the boundary limit for configuration database size?
    2. How we can scale the configuration database?
    3. Can we add multiple configuration database per farm?
    4. What are the best practices and regular maintenance activities to be done to keep the configuration db size under control?
    It would be really really great if someone gives some good knowledge in this area.

    1) Not aware of one
    2) You can add multiple files to the file group
    3) No
    Is the growth in the MDF or the LDF (log file)? If it is in the log file, are you using high availability (Clustering, Mirroring, Log Shipping, AlwaysOn)? If so, you need to run a BACKUP LOG periodically to maintain the LDF size, or allow you to truncate
    the LDF.
    Trevor Seward
    Follow or contact me at...
    &nbsp&nbsp
    This post is my own opinion and does not necessarily reflect the opinion or view of Microsoft, its employees, or other MVPs.

  • Can i use "Oracle Database 12c: Performance Management and Tuning " training for getting certification on "Oracle Database 11g: Performance Tuning 1Z0-054"

    i have taken "Oracle Database 12c: Performance Management and Tuning new" training from oracle university. Now i would like to get certified on "Oracle Database 11g: Performance Tuning 1Z0-054" exam. Is it possible ?

    I essentially endorse and refer you to Matthews' and John's post above.
    I would differ with slightly with Matthew because my guess is you would often be able to use like for like 12c training for an 11g certification ( I believe there are precedents).  BEFORE ANYONE ASKS THE OTHER WAY DOESN'T HAPPEN.
    .... but totally concur with Matthew you would ill advised to procede on that basis without one of:
    - This being advertised as possible on the website : e.g. https://education.oracle.com/pls/web_prod-plq-dad/db_pages.getpage?page_id=654&get_params=p_id:169 ... option 2 show courses.
    - Confirmation from Brandye
    - Confirmation from an Oracle Certification Support Web Ticket ( http://education.oracle.com/pls/eval-eddap-dcd/OU_SUPPORT_OCP.home?p_source=OCP )
    ... The more common (and in my opinion usually better) way would be get your 11g DBA OCP ( or higher first) and then take the 1z0-054.  I am almost certain they will accept your 12c course for the 11g DBA OCP.
    If you are choosing the route of not being a 11g (or 12c ) DBA OCP first but are on option 2 and relying on the course for certification then the issue is more in the balance and you are even more strongly advised to get confirmation before proceding (remember if the rules need to be changed for you only then any profit out of the exam is lost).
    In general my understanding is Oracle would prefer to encourage people to train on the latest version of product that is available for training  and will prefer to avoid restrictions which would cause you to train at a lower version.  ( This is simply my guess at Oracle University Policy ... personal opinion only).
    Having said all I have said I'd encourage you to go with the advice of the earlier two posts.

  • Best Trategy to reduce the Database Size

    Hi Everyone,
    In our Client's Landscape SAP systems have been upgraded to newer versions whereas our client want
    one copy of older Production systems (one copy to retain)
    1) SAP R/3 4.6 C system  (database size of this system is approx 2TB)
    2) SAP BW 3.0 (database size of this system is approx 2TB)
    Now CLient wants us to reduce the database size via re-organization because Archiving of IDOCs & Links we have already done
    Client has recommended for :
    1) Oracle Export/Import: Only Oracle DBA can do (ignore this one)
    2) Database Reorganization : We have tried Reoragnization via BRtools but found very tedious 9 (ignore this one)
    3) SAP Export/Import : Via this way we want to reduce the database size
    Can anybody Tell us How much Free space do we require in order at OS level in order to store the Database Export
    of Two Databases of size around 4TB & what would be the best strategy of reducing the Dabase size.
    Via SAP Export/Import how much approx how much database size will be reduced
    Thanks & Regards
    Deepak Gosain

    Hi,
    >Can anybody Tell us How much Free space do we require in order at OS level in order to store the Database Export
    >of Two Databases of size around 4TB & what would be the best strategy of reducing the Dabase size.
    The only realistic way to know is to do a system copy of the production system on a testbed system and to test the Database Export.
    If you really want to decrease the database size you will have to archive a lot more than the IDOC archiving object.
    Regards,
    Olivier

  • Best practice on mailbox database size & we need how many server for deployment exchange server 2013

    Dear all,
    We have  an server that runs Microsoft exchange server 2007 with the following specification:
    4 servers: Hub&CAS1 & Hub&CAS2 & Mailbox1 & Mailbox2 
    Operating System : Microsoft Windows Server 2003 R2 Enterprise x64
    6 mailbox databases
    1500 Mailboxes
    We need to upgrade our exchange server from 2007 to 2013 to fulfill the following requirment:
    I want to upgrade the exchange server 2007 to exchange server 2013 and implement the following details:
    1500 mailboxes
    10GB or 15GB mailbox quota for each user
    How many
    servers and databases  are
    required for this migration<ins cite="mailto:Mohammad%20Ashouri" datetime="2014-05-18T22:41"></ins>?
    Number of the servers:
    Number of the databases:
    Size of each database:
    Many thanks.

    You will also need to check server role requirement in exchange 2013. Please go through this link to calculate the server role requirement : http://blogs.technet.com/b/exchange/archive/2013/05/14/released-exchange-2013-server-role-requirements-calculator.aspx
    2TB is recommended maximum database size for Exchange 2013 databases.
    Here is the complete checklist to upgrade from exchange 2007 to 2013 : http://technet.microsoft.com/en-us/library/ff805032%28v=exchg.150%29.aspx
    Meanwhile, to reduce the risks and time consumed during the completion of migration process, you can have a look at this proficient application(http://www.exchangemigrationtool.com/) that would also be
    a good approach for 1500 users. It will help you to ensure the data security during the migration between exchange 2007 and 2013.

  • One of the BizTalk Server processes in the affected computer is being throttled for significant periods because of high database size exceeding the threshold

    Hello Experts,
    I have complex Biz Talk 2013 farm having 20 servers,15 Hosts. In my production environment even if there is no traffic i am getting throttling errors from SCOM for my all the production hosts.
    Error : One of the BizTalk Server processes in the affected computer is being throttled for significant
    periods because of high database size exceeding the threshold
     I checked following things:
    1. MsgBoxDB size 748732 KB
    2. Spool table size  53 MB
    3. Tracking DB size 26724 KB
    4. Host settings --  Message Queue Size = 100, MsgCount = 50000 , Spool MP = 10 , Tracking MP = 10
    5. Ran Message Box viewer  and did not found any error related to DB size. (which counter i should focus on in MBV)
    Note -- for DB i am sharing full backup size because it does not have log file size.
    Please suggest where i should focus?
    Is SCOM reporting correctly because everything is fine in biztalk ?
    Thanks
    Yagya
    https://www.mcpvirtualbusinesscard.com/VBCServer/card.aspx?tag=YagyaDattMishra&wa=wsignin1.0

    Hi Yagyam
    I remember this error from SCOM when you use the standard SCOM BizTalk pack.
    Check the eventlog of the server, do you see any errors in event log, this could give some clue to the root cause. Whenever you have this alert from SCOM, you must have
    some entries in eventlog relating to the alert raised by SCOM.
    Is your message processing by BizTalk hosts are normal? Run the performance monitor to see the bottleneck. Check if there is really throttling.
    As mentioned in this blog, check the state of the SQL Agent on BizTalk's SQL database server, may be try restart the SQL Agent on BizTalk's SQL database server.
    http://blogs.msdn.com/b/timdel/archive/2008/11/19/why-i-love-scom.aspx
    If this answers your question please mark it accordingly. If this post is helpful, please vote as helpful by clicking the upward arrow mark next to my reply.

  • How  to increase the database size

    Hi All ,
    Our Client is going to increase their customers from 3.5 million to 5 millions
    (12 millions in future ) and there is no increase in the existing licenses..
    At present we have very less space in Data base server. We are Going Live at the End of the Month.
    How to increase the database size to meet the Client requirements.
    we have to increase the No of Processerors etc...
    Below are the details of Servers:
    System Model: IBM,7040-681
    Machine Serial Number: 830D5BF
    Processor Type: PowerPC_POWER4
    Number Of Processors: 2
    Processor Clock Speed: 1904 MHz
    CPU Type: 64-bit
    Kernel Type: 64-bit
    LPAR Info: 4 bgl-svr-biap
    Memory Size: 4096 MB
    Good Memory Size: 4096 MB
    Note : Server is IBM P690.
    1)  System Lansdcape --> Two System Landsacpe ( Dev to Prod )
    2)  BW Version --> 3.5
    3)  SAP Basis --- > 6.4
    4)  No of Application servers currently having --> Only  One  Server
    Can any body suggest me how to solve this.....
    Thanks in Advance..
    Kasi Chukka

    Hi Kasi,
    Your question is not very clear, from the looks of it, the load on your client's system will be doubling soon.
    You need to look at the system stats on how the system is handling the current load, use transactions ST06, ST04 and ST02 to check if the DB hit ratios are as expected, the cpu idle time, the busy periods. Identify if there is large amounts of paging on the server(candidate for increase in real memory).
    On AIX level, you will need to look at CPU / Memory usage.
    On SAP Application server, you need to look at SAP Buffers.
    On Oracle end, look at SGA stats, if it need to increase, the data growth that is current and estimated growth (DB02).
    You will not get an exact answer to it, you will need to analyze the existing load and then estimate (guess). The hardware upgrade will also depend on what the client is willing to spend, its always a good practise to add 20% to what you have estimated as these figures are never accurate.
    All the best.
    Cheers,
    Nisch

  • How to know the database size

    can one tell how to find the database size?

    Depends what you mean by database size and how you want the output to look?
    Are you referring to the number of rows across all tables of a schema? The number of bytes used by the rows? The number of bytes used by the tablespaces? The number of bytes used by the datafiles? etc. etc. etc. So many options.
    Be more clear in your requirements please.

Maybe you are looking for

  • How to delete just one page in Pages 4.3

    Hello I am having an issue with trying to delete just one page out of the 47 pages that I have created. If I try to delete page 12 it will delete 12-47 instead of just page 12. I also cant seem to move or adjust the yellow box in thumnails in order t

  • What's wrong with this code (in recursion there may be a problem )

    // ELEMENTS OF TWO VECTORS ARE COMPARED HERE // CHECK WHAT IS WRONG WITH THIS PGM // there may be a problem in recursion import java.util.*; class Test Vector phy,db; public static void main(String adf[])      Vector pp1=new Vector();      Vector dp1

  • Has anyone else had a problem with the screen cracking on the S5?

    I have NOT dropped my phone, yet the screen cracked in two places. The LCD is fine.  I called Verizon, who told me to call Samsung. I called Samsung, on Verizon's advice, and was told they are aware of the problem. I can send it back to Samsung and t

  • Putting comments from Planning in a report via SQL

    Has anyone built a report in Web Analysis using the SQL Spreadsheet to report comments entered in via the Planning Application. I found three tables in the Planning repository that contain notes, two of them tie together but the third is too generic

  • A Problem with IBlobFieldObject

    Hi all, I am using Crystal Reports version 10.2. I used to work with Crystal Reports XI release 2, but I downgraded it because my application is 64 bit. When i used the newer version I didn't have any problems loading image files into IBlobFieldObjec