How define database performance?

Hi all
I have a question about performance.
I have an Oracle database installed on Xeon server
1. I want to know parameters of my server. How many processors it has? What is it's RAM and other parameters. From Linux, with which command I can get information about it?
2. My company wants to buy a server computer for our database. My questions is: Depending on which parameters of database, which server I must select??
Thanks...

All my vmstat is like that
vmstat 10 13
procs -----------memory---------- ---swap-- -----io---- system ----cpu----
r b swpd free buff cache si so bi bo in cs us sy id wa
1 3 78696 16300 3836 2546896 23 19 1447 62 220 106 56 12 19 13
2 3 91188 15468 3852 2535148 2 128 5285 195 2223 2231 31 8 24 37
2 6 124956 16620 3776 2501260 0 286 8264 422 2228 2220 30 9 13 47
2 3 141692 20268 4120 2493528 0 1467 4935 1589 2266 2233 42 11 12 35
3 2 153172 17516 4380 2494264 0 1098 3174 1216 2051 2081 48 9 18 26
2 2 158496 20188 4688 2492288 16 631 3305 754 2081 2318 42 9 20 29
2 2 167652 24556 5032 2494788 4 642 3353 782 2023 2408 47 8 19 26
11 1 171272 15660 5556 2505372 13 659 2998 827 1972 2049 66 9 10 15
0 2 175784 17452 5928 2511704 26 342 2632 435 1958 1893 63 10 14 12
3 1 184056 18156 5976 2504476 0 889 2139 1027 1842 1783 64 9 14 12
5 4 185196 16988 6352 2510060 13 567 2192 992 1787 1652 73 8 8 11
7 2 185324 24748 6660 2503840 0 1146 2831 1332 1858 1663 74 11 5 11
10 2 185308 25700 7132 2507532 0 77 1976 202 1901 1297 85 14 0 1

Similar Messages

  • How to report database performance to upper management?

    Hello all. 
    I am looking for a way to report database performance to upper management.  I'll start by giving you an overview of our team structure. I manage a team of various roles: Level 2 support specialists, configuration managers, and DBAs - we are dubbed the
    "Production Support Team".  I report directly to the VP of Software Development who oversees the Design, DEV, and QA, team managers.  This means that our team, Production Support, can be considered a peer team of Design, DEV, and QA. 
    The manager of each team compiles metrics from their team members and reviews them on a weekly basis to ensure we are headed in the right direction.  We then take our metrics, and distill them down into a few key metrics that we then report to the VP
    of Software Development.  Those metrics are easy tom compile for the L2 support roles (number of tickets escalated, number closed, etc.).  However, I've been struggling with how to report DBA performance for a long time.  Every time I start
    to compile numbers, I end up with tons of things that are great for a DBA, but don't mean much to upper management.  Does anyone have any suggestions.  Surely I'm not the only person that has had this issue?
    I'd be happy to provide more info if needed.  Thanks!
    Clint

    are you using an RTF template ?
    in the RTF template you need to use this syntax for the HTML enabled field
    <?html2fo:MOSC_COMMENTS?>
    instead of <?MOSC_COMMENTS?>

  • How can i define database name in oracle ADI

    Hai Friends,
    I am facing a problem when i am going to define database name in oracle ADI there is a error u can enter minimum 6 char database name but my database name is PROD it is only 4 char long. how can i define can i will change the database name. Please suggest me.
    If u have any query regarding oracle apps plse discuss with me.
    Many thanks
    Ghanshyam khetan

    Database is used internally by ADI and does not have any relationship to the name of your database.
    Using ADI I am able to create a database with the following:
    Name - My Wacky Database
    GWYUID and FNDNAM as usual
    Connect String - VIS
    I'm able to signon and work successfully with that definition. The connect string is more important in this scenario as that tells the networking layer which database to actually talk to. As you can see here mine is only three characters.

  • How about  a separate forum Database-Performance

    hi friends/Moderators,
    I had one thing to give feedback/suggestion to give , but didnt know of any link/place to put it.
    So opening a new thread over it.
    I see many new threads related to Database Performance, sql tuning, redo/undo ( and many other architecture related things which can impact the performance of the database).
    How about a idea of having another forum Database-Performance like we have Database-General now?
    Personally I would like to start replying to performance related questions so that apart from helping the OPs i learn more about performance issues which other people see or face outside of my workplace.
    What do you guys say?
    Cheers!

    Kunwar wrote:
    hi friends/Moderators,
    I had one thing to give feedback/suggestion to give , but didnt know of any link/place to put it.
    So opening a new thread over it.For such questions, please post them over the Community Discussions forum.
    I see many new threads related to Database Performance, sql tuning, redo/undo ( and many other architecture related things which can impact the performance of the database).
    How about a idea of having another forum Database-Performance like we have Database-General now?
    Personally I would like to start replying to performance related questions so that apart from helping the OPs i learn more about performance issues which other people see or face outside of my workplace.
    What do you guys say?
    This was asked by me in 2008. Please have a read of the thread,
    Can We Have a Perfomance Tuning Forum?
    That said, it may break things even more. Don't take it personally but just take the example of this question. You are here from some time now and even then, you wrongly posted this question in a product forum even when there is already a non-technical, community feedback forum is in place. So may be people won't use a separate forum of Performance and will keep on posting the questions in different forums, just like what happens now too. And with every new forum added, the overhead would also increase (my guess) .
    And another thing that comes to mind is that since the forums are already scheduled for a major upgrade in the mid-April , I doubt that such request would be fulfilled , at least right now because last time, when the upgrade of the forums happened, it wasn't a smooth one and the focus of the OTN team would be surely to make sure that the history won't repeat itself.
    Just my 2 cents.
    Aman....

  • Database Performance Checks

    Oracle : 10.2.0.3
    OS : Linux 64 bit
    Issue : Slow performance at 11-30PM complained by client.
    Checks done :
    1. Ran AWR between 11 PM and 12 Noon.
    CPUs : 4 SGA Size : 2,000M (100%) Buffer Cache : 1,584M (79.2%) Shared Pool 1,129M (56.4%)
    ADDM suggest SGA_TARGET to increase from 2000MB to 2500MB.
    2. top 5 events
    Top 5 Timed Events                                         Avg %Total
    ~~~~~~~~~~~~~~~~~~                                        wait   Call
    Event                                 Waits    Time (s)   (ms)   Time Wait Class
    db file scattered read            1,952,811       4,804      2   30.5   User I/O
    CPU time                                          3,448          21.9          
    db file sequential read             149,712       1,921     13   12.2   User I/O
    read by other session               293,022         877      3    5.6   User I/O
    log file sync                         9,920         157     16    1.0     Commit
              -------------------------------------------------------------       3. Stats are upto date.
    4. Index rebuild requirement is not there
    SQL> SELECT name,height,lf_rows,del_lf_rows,(del_lf_rows/lf_rows)*100 as ratio FROM INDEX_STATS;
    no rows selected5. Average 100 sessions will connect to the database
    6. Checked all logs fr any disconnection details
    7. Application is running from weblogic
    Questions : How to certify the performance is good or slow from the above observations. I am able to feel the statistics are the similar for the different periods where I ran AWR report.
    : Other than the user as a DBA what are the other checks can be done to monitor the performance

    It's difficult to use AWR or Statspack to "certify" database performance is good. It just depends what "performance is good" means.
    Most of the time it's application response time which is the right metric: database response time is only a part of application response time and AWR/Statspack cannot easily link database response time and application response time.
    [11.2 Concepts Guide Principles of Application Design and Tuning| http://download.oracle.com/docs/cd/E11882_01/server.112/e10713/cncptdev.htm#CHDEHHIJ] says
    >
    Define clear performance goals and keep historical records of metrics
    An important facet of development is determining exactly how the application is expected to perform and scale. For example, you should use metrics that include expected user load, transactions per second, acceptable response times, and so on. Good practice dictates that you maintain historical records of performance metrics. In this way, you can monitor performance proactively and reactively (see "Performance Diagnostics and Tuning").

  • How can I perform this kind of range join query using DPL?

    How can I perform this kind of range join query using DPL?
    SELECT * from t where 1<=t.a<=2 and 3<=t.b<=5
    In this pdf : http://www.oracle.com/technology/products/berkeley-db/pdf/performing%20queries%20in%20oracle%20berkeley%20db%20java%20edition.pdf,
    It shows how to perform "Two equality-conditions query on a single primary database" just like SELECT * FROM tab WHERE col1 = A AND col2 = B using entity join class, but it does not give a solution about the range join query.

    I'm sorry, I think I've misled you. I suggested that you perform two queries and then take the intersection of the results. You could do this, but the solution to your query is much simpler. I'll correct my previous message.
    Your query is very simple to implement. You should perform the first part of query to get a cursor on the index for 'a' for the "1<=t.a<=2" part. Then simply iterate over that cursor, and process the entities where the "3<=t.b<=5" expression is true. You don't need a second index (on 'b') or another cursor.
    This is called "filtering" because you're iterating through entities that you obtain from one index, and selecting some entities for processing and discarding others. The white paper you mentioned has an example of filtering in combination with the use of an index.
    An alternative is to reverse the procedure above: use the index for 'b' to get a cursor for the "3<=t.b<=5" part of the query, then iterate and filter the results based on the "1<=t.a<=2" expression.
    If you're concerned about efficiency, you can choose the index (i.e., choose which of these two alternatives to implement) based on which part of the query you believe will return the smallest number of results. The less entities read, the faster the query.
    Contrary to what I said earlier, taking the intersection of two queries that are ANDed doesn't make sense -- filtering is the better solution. However, taking the union of two queries does make sense, when the queries are ORed. Sorry for the confusion.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Regarding Database Performance

    Hi All,
    I have installed *10gR2 on RHEL4 (4GB -- RAM, space is enough)*. One application (oracle ucm) is running on that. Its contains apache and content server. After 2-3 weeks, developers were saying taking long time for opening url. So done gather database statistics (after that daily gathering db stats using scheduler). After that, it was working fine. Again after week they are having the prob. They are doing lot of dml on db. Checked in os level using top command. But oracle ( installed entire application as oracle) user is not consuming that much memory. set pga_aggregate_target to about 500M. Sga (sga_max_size --- 950M) is auto tuning. db is of size 8GB. workarea_policy_size is auto.
    Please suggest any solutions for improving database performance.
    Thanks,
    Manikandan.

    daily gathering db stats using scheduler)Done by default on V10+
    Please suggest any solutions for improving database performance.Ready, Fire, Aim!
    Is any OS resource the bottleneck; CPU, RAM, IO, network?
    During slow period what is reported by AWR?
    Please read these:
    When your query takes too long
    When your query takes too long ...
    How to Post a SQL statement tuning request
    HOW TO: Post a SQL statement tuning request - template posting
    Edited by: sb92075 on Jul 27, 2010 10:01 AM

  • Oracle 11 G database performance tuning

    How to indexing the oracle 11G database?
    Thanks in advance.

    Your question is like if you ask "Tell me how to fix a car". As you know people spend years learning how to fix different problems with cars; the same applies to database performance tuning. There is no way to answer this question in one post. Please ask a more specific question in an apropriate (database related) forum.
    cheers

  • Database Performance Assesment

    Hi,
    is there any way to asses the performance of my database?

    In addition, what you can do is to build a periodic profile (just like a detective) of your database activities at specified times for your database. this can be done using Statspack or AWR depending on your database release.
    For instance, If your customers mostly access the database application between 8 am and 5 pm, you can take snapshots of your database activities at peak times, off peak, during heavy transaction loads or at any specified date and time. Then you can compare those snapshots to give you an idea of how your database is performing at those different times. This snapshots can be set to be taken automatically every 30 minutes or 1 hour and also taken at any adhoc time you want.
    This way, you can come up with a Management report/graph of the load/Performance of your database for different times of the day, different times of the week, different times of the month, different load type and different customer activity times etc

  • Database Performance Evaluation Benchmarking Tuning

    Does anyone by chance have any articles or websites that deal with Oracle (or generic) database performance evaluation, benchmarking, and/or tuning for a STANDALONE PC database installation? I have Oracle 10g installed on my personal machine and I want to find some information that will help me place the performance evaluation into a more professinal context. Any links relating to this would be very appreciated.
    regards,
    John

    Why do you expect there to be a difference in evaluating performance for a standalone PC database installation vs. a server database installation? Other than the fact that, presumably, you won't find network events in the top wait events, I'm hard-pressed to think of any differences.
    You would evaluate performance and tune the system the same way on a standalone database as on a database on a server box-- you would figure out what operations are important, figure out how quickly those operations need to run, figure out which of the important operations are running slowly, figure out what those operations are spending time doing, and then figure out how to reduce the runtime. Of course, each of those steps can potentially be rather involved. There are plenty of articles and books on performance tuning-- Oracle has a few manuals, Jonathan Lewis's book on the cost based optimizer is excellent, Cary Millsap's optimization book is top flight, etc.
    As for benchmarking, unless the intention is to run something like the TPC benchmark on your desktop, which would seem odd, your benchmark is generally closely tied to your application-- i.e. figuring out how quickly the system performs a particular business operation. Generic benchmarks like TPC tend not to be particularly useful in the real world because they are unlikely to mimic your real workload.
    Justin

  • How to do performance tuning in EXadata X4 environment?

    Hi,  I am pretty new to exadata X4 and we had a database (oltp /load mixed) created and data loaded. 
    Now the application is testing against this database on exadata.
    However they claimed the test results were slower than current produciton environment. and they send out the explain plan, etc.
    I would like advices from pros here what are specific exadata tuning technics I can perform to find out why this is happening.
    Thanks a bunch.
    db version is 11.2.0.4

    Hi 9233598 -
    Database tuning on Exadata is still much the same as on any Oracle database - you should just make sure you are incorporating the Exadata specific features and best practice as applicable. Reference MOS note: Oracle Exadata Best Practices (Doc ID 757552.1) to help configuring Exadata according to the Oracle documented best practices.
    When comparing test results with you current production system drill down into specific test cases running specific SQL that is identified as running slower on the Exadata than the non-Exadata environment. You need to determine what is it specifically that is running slower on the Exadata environment and why. This may also turn into a review of the Exadata and non-Exadata architecture. How is application connected to the database in the non-Exadata vs Exadata environment - what's the differences, if any, in the network architecture in between and the application layer?
    You mention they sent the explain plans. Looking at the actual execution plans, not just the explain plans, is a good place to start... to identify what the difference is in the database execution between the environments. Make sure you have the execution plans of both environments to compare. I recommend using the Real Time SQL Monitor tool - access it through EM GC/CC from the performance page or using the dbms_sql_tune package. Execute the comparison SQL and use the RSM reports on both environments to help verify you have accurate statistics, where the bottlenecks are and help to understand why you are getting the performance you are and what can be done to improve it. Depending on the SQL being performed and what type of workload any specific statement is doing (OLTP vs Batch/DW) you may need to look into tuning to encourage Exadata smart scans and using parallelism to help.
    The SGA and PGA need to be sized appropriately... depending on your environment and workload, and how these were sized previously, your SGA may be sized too big. Often the SGA sizes do not usually need to be as big on Exadata - this is especially true on DW type workloads. DW workload should rarely need an SGA sized over 16GB. Alternatively, PGA sizes may need to be increased. But this all depends on evaluating your environment. Use the AWR to understand what's going on... however, be aware that the memory advisors in AWR - specifically for SGA and buffer cache size - are not specific to Exadata and can be misleading as to the recommended size. Too large of SGA will discourage direct path reads and thus, smart scans - and depending on the statement and the data being returned it may be better to smart scan than a mix of data being returned from the buffer_cache and disk.
    You also likely need to evaluate your indexes and indexing strategy on Exadata. You still need indexes on Exadata - but many indexes may no longer be needed and may need to be removed. For the most part you only need PK/FK indexes and true "OLTP" based indexes on Exadata. Others may be slowing you down, because they avoid taking advantage of the Exadata storage offloading features.
    You also may want evaluate and determine whether to enable other features that can help performance including configuring huge pages at the OS and DB levels (see MOS notes: 401749.1, 361323.1 and 1392497.1) and write-back caching (see MOS note: 1500257.1).
    I would also recommend installing the Exadata plugins into your EM CC/GC environment. These can help drill into the Exadata storage cells and see how things are performing at that layer. You can also look up and understand the cellcli interface to do this from command line - but the EM interface does make things easier and more visible. Are you consolidating databases on Exadata? If so, you should look into enabling IORM. You also probably want to at least enable and set an IORM objective - matching your workload - even with just one database on the Exadata.
    I don't know your current production environment infrastructure, but I will say that if things are configured correctly OLTP transactions on Exadata should usually be faster or at least comparable - though there are systems that can match and exceed Exadata performance for OLTP operations just by "out powering" it from a hardware perspective. For DW operations Exadata should outperform any "relatively" hardware comparable non-Exadata system. The Exadata storage offloading features should allow you to run these type of workloads faster - usually significantly so.
    Hope this helps.
    -Kasey

  • How to Achieve Performance Tuning In BPM Studio

    Please Tell me how to achieve performance Tuning in BPm Project . let me know do have any documentation for this .
    Thanks in Advance .

    *5. Group Automatic Activities in a Single Transactional Boundary*
    When you have several automatic activities in a sequence, recognize this as a potential performance improvement opportunity. The default behavior of Oracle BPM is during each Automatic activity's execution:
    1. Initiate the transaction
    2. Read the work item instance's variable information from the Engine's database
    3. Execute the logic in the Automatic activity
    4. If no system exception occurs, commit the transaction and write the instance variable information back to the Engine's database
    Many times you'll instead want to speed execution when there are several Automatic activities in a sequence. If three Automatic activities are in a sequence, then the four items listed above will occur three times. By grouping these into a single transactional boundary, instead of 12 steps you would have:
    1. Initiate the transaction
    2. Read the work item instance's variable information from the Engine's database
    3. Execute the logic in the first Automatic activity
    4. Execute the logic in the second Automatic activity
    5. Execute the logic in the third Automatic activity
    6. If no system exception occurs, commit the transaction and write the instance variable information back to the Engine's database
    This grouping of Automatic activities into a single transactional boundary can be done in one of these three ways:
    1. Create a Group around the sequence of Automatic activities (lasso the three activities) -> right mouse click inside the dotted line -> click "Create a Group with Selection" -> click "Runtime" in the upper left corner -> click the checkbox "Is Atomic".
    2. Instead of placing the Automatic actiivities in the process, add them in a Procedure and then call the Procedure from a new Automatic activity in the process.
    3. In Oracle BPM 10g you can enable "Greedy" execution for the process by right mouse clicking the process's name in the Project Navigator tab -> click "Properties" -> click the "Advanced" tab -> click the "Enable Greedy Execution" radio button.
    Dan

  • How to do Performance analysis of any ABAP report

    Hello Friends,
                         I have to do the performance analysis of ABAP report, it takes time to execute. I am  not aware about, how to do performance analysis? I know only we have check index, do ST04 and ST05 analysis. but how to do these things  that i don't know.
                        Can anyone please guide me in this regards.
    Thanks

    HI,
    Please Check this,
    System Trace: Transaction ST01 lets you do various levels of system trace such as authorization checks, SQL traces, table/buffer trace etc. It is a general Basis tool but can be leveraged for BW.
    Workload Analysis: You use transaction code ST03
    Database Performance Analysis: Transaction ST04 gives you all that you need to know about whatu2019s happening at the database level.
    Performance Analysis: Transaction ST05 enables you to do performance traces in different are as namely SQL trace, Enqueue trace, RFC trace and buffer trace.
    ABAP Runtime Analysis Tool: Use transaction SE30 to do a runtime analysis of a transaction, program or function module. It is a very helpful tool if you know the program or routine that you suspect is causing a performance bottleneck.
    Rgds
    Sabu

  • Database performance issue (8.1.7.0)

    Hi,
    We are having tablespace "payin" in our database (8.1.7.0) .
    This tablespace is the main Tablespace of our database which is dictionary managed and heavily accessed by the user SQL statements.
    Now we are facing the database performance issue during the peak time (i.e. at the month end) when no. of users use to run the no. of large reports.
    We have also increased the SGA sufficiently on the basis of RAM size.
    This tablespace is heavily accessed for the reports.
    Now my question is,
    Is this performance issue is because the tablespace is "dictionary managed" instead of locally managed ?
    because when i monitor the different sessions through OEM, the no. of hard parses is more for the connected users.
    Actually the hard parses should be less.
    In oracle 8.1.7.0 Can we convert dictionary managed tablespace to locally managed tablespace ?
    by doing so will the problem will get somewhat resolve ? will it reduce the overhead on the dictionary tables and on the shared memory ?
    If yes then how what is procedure to convert the tablespace from dictionary to locally managed ?
    With Regards

    If your end users are just running reports against this tablespace, I don't think that the tablespace management (LM/DM) matters here. You should be concerned more about the TEMP tablespace (for heavy sort operations) and your shared pool size (as you have seen hard parses go up).
    As already stated, get statspack running and also try tracing user sessions with wait events. Might give you more clues.

  • New HP EVA6000 SAN and now bad database performance problems

    Hello,
    we changed our SAN Hardware to a HP EVA6000 and moved all Data to there.
    It is intended that storagesystem is for all Servers (File/Print, Exchange, Oracle Databases and MSSQL Databases).
    According to the best practice papers of HP we created one big discgroup (fata harddiscs) and created virtual discs for our servers.
    After doing this the database performance was terrible bad!
    Especially the multiple random IO is absolutely worse.
    As a first countermeasure we created a second discgroup on faster harddiscs and moved the database contents to there. We analyzed the IO and moved several database file to different virtual discs.
    The performance is better now, but still not like the 4 year old SAN System!
    Of course we questioned HP and even let them do a performance analysis but up to now we have no solution... The performance analysis report will be available on thursday.
    Does anybody had the same experience or how did you configure the Database and EVA SAN to work with an appropiate performance?

    Hi,
    I'm not an Oracle person, but do work with EVA SAN's.
    Your 48 FATA disk, disk group is capable of 4800 I/O operations per second [48 x 100] but the disks are only rated for 30% duty cycle and spin at 7200 rpm.
    The 16 FC drive Disk Group I/O operations rating depends on the speed the disks spin at. 10K rpm disks are rated at 130 I/O per sec [2080 for the group] and if they are 15K rpm disks then 170 I/O per sec [2720 for the group]. Both are rated for 100% duty cycle.
    I seem to recall having have read somewhere that Oracle prefers to have it's logs on separate storage to it's data.
    If your shelves had the spare disk slots I would put in 72 GB 15K rpm disks up to the capacity required {+ overhead } + head room for reasonably predictable growth over the anticpated life of the equipment.
    Here is a link to the HP Best Practice guide for EVA 4/6/8000's
    http://h71028.www7.hp.com/ERC/downloads/4AA0-2787ENW.pdf
    I hope this helps you understand the storage that you are working with a bit better. The old saying of "more heads make for better performance" is still true, however budget can have some affect in performance. ;-)
    Jim

Maybe you are looking for

  • Opening and closing folders very difficult while organizing projects

    Alo, after importing my iPhoto Library to Aperture, I'm working through the mass list of iPhoto Events, now projects. Each time I move a project from the iPhoto Library folder, to a folder that I've created, the very quick spring loaded action takes

  • Is there a method of adding a privacy feature to one or more of my bookmarks?

    I would like to know if there is any way to password protect or add a privacy feature to certain of my bookmarks.

  • E71 Doesnt Boot Up Properly

    Hi Everyone, My brand new E71 doesn't boot up anymore. I'm guess it is because of "PowerBoot" applications or whatever apps it is loading. After I've installed a couple of applications today it was all fine, until I did a reboot. Now when I boot it u

  • Sql Developer Data Modeler 3.0 EA1: Export - To reporting schema

    When launching Export -> To reporting schema the schema tables are created and messages show that data is being generated but after indicating that it "Exporting Diagrams" I get the following error message: "Error occurred while exporting to Reportin

  • How to create a button Widget

    Hi, Can any one tell me how to create a button widget, I've created one & it can be seen in my plugin, but it doesn't work, in .fr file i created the widget as, ButtonWidget            kEXTCODGoButtonWidgetID,                  kSysButtonPMRsrcId,