Reducing the memory utilisation of my database

Hi,
I want to reduce the memory utilisation of my database. I want to know which sqls have assigned by the oracle some OS memory in my database.I have the awr reports with me.
My questions:
1.Which section of the awr will give exactly this information?
(SQL ordered by Sharable Memory doesn't help )
2. Or can you tell me some views or tables wherein I get the needed the information which I can query against in my database?
3. How can I reduce the memory utilisation in case I get the problematic sqls?
Thanks,
Sach

I'm not sure that I understand your question. Can you clarify a couple points for me?
What memory are we talking about here? Normally, most of the RAM allocated to Oracle is going to be SGA. But SGA isn't associated with any particular SQL statement, at least not in a fashion that I could contemplate doing reporting on. Individual SQL statements require RAM temporarily in the PGA during execution, but it sounds like you're not interested in that.
What is the problem you are trying to solve here? If you want to reduce the amount of RAM allocated to Oracle from the operating system, you should be able to do that without analyzing any specific SQL statements by adjusting memory parameters. Mentioning what version of Oracle, what parameters you've set, and how much you'd like to reduce memory consumption would be helpful if you want specific suggestions for parameters to change.
What does "problematic sqls" mean in this context?
Justin

Similar Messages

  • Reducing the memory footprint of our Sybase ASE based SolMan install

    Hello All,
    We are doing a test install of SAP Solution Manager 7.01 on Sybase ASE 15.7.
    Since this is just a test setup, we started off with a lower-than-recommended hardware configuration (4 GB RAM only) due to time constraints and since we were 'assured' that we could do basic testing with this setup.
    While post install performance of SolMan was decent, performance during solman_setup (setting up technical monitoring) has become appalling. We are not able to complete the configuration process at all as the SolMan configuration web application has become very unpredictable and extremely slow.
    The SolMan install is centralized and on a windows 2008 box. Windows task manager shows consistent memory usage of up to 90 - 95%. We also tried reducing the total number of work processes to just 8 but that did not help much. We see in 'task manager > resource monitor' that sqlserver.exe process is taking a shareable working set close to 2 GB of RAM whereas the committed memory much less (34 MB). Please tell us about any memory optimization we can perform for SolMan / Sybase ASE in order to complete technical monitoring setup using Solman_setup. We were hoping that we could change the  'total logical memory' setting for the DB directly using DBACOCKPIT tcode (in order to reduce the max memory setting) but could not do so as the it seems to be read-only. We could not find much documentation/posts regarding memory optimization for the DB. Please help out. Thanks!
    -Regards,
    Arvind

    FWIW ... ASE's 'max memory' setting can be changed on the fly, while 'total logical memory' is a calculated value that you cannot change (ie, it's 'read only'; changing 'max memory' will cause 'total logical memory' to change automatically). [NOTE: DBACOCKPIT is a SAP-provided application that sits on top of ASE; while I know what's doable when connected directly to ASE I do not know if DBACOCKPIT has disabled the ability to change some configuration settings like 'max memory'.]
    As for the SolMan performance issues ... I'd recommend reposting your issue in the SAP Applications on ASE discussion group where you're likely to get the attention of more folks with SAP application (on ASE) experience.  (While someone may jump in here with SolMan suggestions, SolMan is a SAP application and this group isn't really geared towards SAP applications.)

  • Is this the best approach to reduce the memory ??

    Hi -
    I have been given a task to reduce the HEAP memory so that the system can support more number of users. I have used various suggestions given in this forum to find out the size of the object in memory. I have reached to a point that where i think i got an approx size of the object in memory.(not 100%)
    I basically have some objects of some other class which are created when this object is created . The intent was to initialize the nested objects once and use them in the main object. I saw some significant difference reduction in size of the object when i create these objects local to the methods which use it.
    Before moving the objects to method level
    Class A {
        Object b = new Object();
        Object c = new Object();
        Object d = new Object();
         public void method1 () {
             b.someMethod();
         public void method2 () {
             b.someMethod();
         public void method3 () {
             c.someMethod();
         public void method4 () {
             c.someMethod();
         public void method5 () {
             d.someMethod();
         public void method6 () {
             d.someMethod();
    After moving the objects to method level
    Class A {
         public void method1 () {
           Object b = new Object();
             b.someMethod();
         public void method2 () {
            Object b = new Object();
             b.someMethod();
         public void method3 () {
           Object c = new Object();
             c.someMethod();
         public void method4 () {
          Object c = new Object();
             c.someMethod();
         public void method5 () {
            Object d = new Object();
             d.someMethod();
         public void method6 () {
            Object d = new Object();
             d.someMethod();
    }Note : This object remains in the http session atleast 2 hrs. I cannot change the session time out.
    Is this the better approach to reduce the heap size? What are the side effects of creating all objects in the local methods which will be on stack?
    Thanks in advance

    The point is not that the objects are on the stack - they aren't, all objects are in heap, but that they have a much shorter life. They'll become unreachable as soon as the method exits, rather than surviving until the session times out. And the garbage collector will probably recycle them pretty promptly, because they remain in "Eden space".
    (In future versions of the JVM Sun is hoping to use "escape analysis" to reclaim such objects even faster).
    Of course some objects might have a significant creation overhead, in which case you might want to consider creating some kind of pool of them from which one could get borrowed for the duration of the call. With simple objects, though, the overheads of pooling are likely to be higher.
    Are these objects modified during use? If not then you might simply be able to create one instance of each for the whole application, and simply change the fields in the original class to static. The decision depends on thread safety.

  • Reducing the memory footprint of Sybase ASE

    Hello All,
    We are doing a test install of SAP Solution Manager 7.01 on Sybase ASE 15.7.
    Since this is just a test setup, we started off with a lower-than-recommended hardware configuration (4 GB RAM only) due to time constraints.
    While post install performance of SolMan was decent, performance during solman_setup (setting up technical monitoring) has become appalling. We are not able to complete the configuration process at all as the SolMan configuration web application has become very unpredictable and extremely slow.
    The SolMan install is centralized and on a windows 2008 box. Windows task manager shows consistent memory usage of up to 90 - 95%. We also tried reducing the total number of work processes to just 8 but that did not help much. We see in 'task manager > resource monitor' that sqlserver.exe process is committing close to 2 GB of RAM when in fact it seems to be using much lesser when seen directly in 'task manager > process tab'. Please tell us about any memory optimization we can perform for Sybase ASE in order to complete out Solman_setup. We were hoping that we could change the  'max memory'/'total logical memory' setting for the DB directly using DBACOCKPIT tcode but could not do so as the parameters seems to be read-only. We could not find much documentation regarding memory optimization for the DB. Please help out. Thanks!
    -Regards,
    Arvind

    FWIW ... ASE's 'max memory' setting can be changed on the fly, while 'total logical memory' is a calculated value that you cannot change (ie, it's 'read only'; changing 'max memory' will cause 'total logical memory' to change automatically). [NOTE: DBACOCKPIT is a SAP-provided application that sits on top of ASE; while I know what's doable when connected directly to ASE I do not know if DBACOCKPIT has disabled the ability to change some configuration settings like 'max memory'.]
    As for the SolMan performance issues ... I'd recommend reposting your issue in the SAP Applications on ASE discussion group where you're likely to get the attention of more folks with SAP application (on ASE) experience.  (While someone may jump in here with SolMan suggestions, SolMan is a SAP application and this group isn't really geared towards SAP applications.)

  • Reducing SQL memory

    Forum,
    What impact will reducing the memory SQL uses have on SAP B1?
    Whilst I appreciate this isn't something that's advised be altered and ideally the more memory the better but I have an instance where there is a 3rd part application and the SQL server is having issues with the about of memory being used thus slowing everything down.
    Regards,

    you can tune the memory parameter according to your situation, Monitor "Page life expectancy" from DB performance history  and make sure this doesn't go down very low. Microsoft recommends below 300 is a problem but its prefer to have this number very high always.
    Thanks
    Mushtaq

  • Will oops concept reduces the allocation space?

    Hi,
    As we know SAP allocates memory for a program by considering it as transaction.
    If i develop my program by using some local classes for a set of code, will this reduce the memory usage of this program?
    --Naveen Inuganti

    Hi,
    As we know SAP allocates memory for a program by considering it as transaction.
    If i develop my program by using some local classes for a set of code, will this reduce the memory usage of this program?
    --Naveen Inuganti

  • Finding Memory Utilisation

    Hi
    I need to find memory utilisation using my java program.??
    Heard about runtime class total memory and free memory but I think it returns that of JVM.
    I dont want JVM's memory utilisation but the memory utilisation of the system.
    Any suggestions for finding the above.

    Guyss...any answers...

  • Best Trategy to reduce the Database Size

    Hi Everyone,
    In our Client's Landscape SAP systems have been upgraded to newer versions whereas our client want
    one copy of older Production systems (one copy to retain)
    1) SAP R/3 4.6 C system  (database size of this system is approx 2TB)
    2) SAP BW 3.0 (database size of this system is approx 2TB)
    Now CLient wants us to reduce the database size via re-organization because Archiving of IDOCs & Links we have already done
    Client has recommended for :
    1) Oracle Export/Import: Only Oracle DBA can do (ignore this one)
    2) Database Reorganization : We have tried Reoragnization via BRtools but found very tedious 9 (ignore this one)
    3) SAP Export/Import : Via this way we want to reduce the database size
    Can anybody Tell us How much Free space do we require in order at OS level in order to store the Database Export
    of Two Databases of size around 4TB & what would be the best strategy of reducing the Dabase size.
    Via SAP Export/Import how much approx how much database size will be reduced
    Thanks & Regards
    Deepak Gosain

    Hi,
    >Can anybody Tell us How much Free space do we require in order at OS level in order to store the Database Export
    >of Two Databases of size around 4TB & what would be the best strategy of reducing the Dabase size.
    The only realistic way to know is to do a system copy of the production system on a testbed system and to test the Database Export.
    If you really want to decrease the database size you will have to archive a lot more than the IDOC archiving object.
    Regards,
    Olivier

  • I need to reduce the size of several pdf files before transferring them onto a research database. Each file is made up of 100 images (pages of diaries)

    I need to reduce the size of several pdf files before transferring them onto a research database. Each file is made up of 100 images (pages of diaries) & is up to 1.36 GB in size. I have tried reducing file size in preview but the images are unreadable. Any ideas? 

    WHen I want to reduce the file size, I either use "Optimize Scanned PDF", or more often use "Reduce File Size". Both are on the Document menu. The Reduce File Size selection has some limited customization that is used, when selected.

  • How to reduce the database size based on MS SQL/Oracle?

    Hi,
    I had deleted severy clients via SCC5, but the database size is more and more bigger. Do you know how to reduce the database size.
    my platform is: NT, MS SQL 2005
                           NT, Oracle 10.2
    Best Regards,
    Simon Shen

    Hi Simon,
    In the case of SQL you need to check how many space free you have in your datafiles and the based on that you need to Shrink them using DBCC SHRINKFILE command.
    Find a couple of reference docs here,
    http://www.intermedia.net/support/kb/default.asp?id=996
    http://doc.ddart.net/mssql/sql70/dbcc_21.htm
    Also i'm pretty sure that if your check the properties of the datafiles in MSSQL Enterprise Manager you get a "Shrink" option.
    Regards
    Juan
    Please reward with points if helpful

  • Query to find the memory of database in oracle,sql

    Hi All,
    Please let me know the query to find the memory of database in oracle,sql.
    Thanks,
    sajith

    How do I find the overall database size?
    The biggest portion of a database's size comes from the datafiles. To find out how many megabytes are allocated to ALL datafiles:
    select sum(bytes)/1024/1024 "Meg" from dba_data_files;
    To get the size of all TEMP files:
    select nvl(sum(bytes),0)/1024/1024 "Meg" from dba_temp_files;
    To get the size of the on-line redo-logs:
    select sum(bytes)/1024/1024 "Meg" from sys.v_$log;
    Putting it all together into a single query:
    select a.data_size+b.temp_size+c.redo_size "total_size"
    from ( select sum(bytes) data_size
    from dba_data_files ) a,
    ( select nvl(sum(bytes),0) temp_size
    from dba_temp_files ) b,
    ( select sum(bytes) redo_size
    from sys.v_$log ) c;
    Another query ("Free space" reports data files free space):
    col "Database Size" format a20
    col "Free space" format a20
    select round(sum(used.bytes) / 1024 / 1024 ) || ' MB' "Database Size"
    , round(free.p / 1024 / 1024) || ' MB' "Free space"
    from (select bytes from v$datafile
    union all
    select bytes from v$tempfile
    union all
    select bytes from v$log) used
    , (select sum(bytes) as p from dba_free_space) free
    group by free.p
    This is what I use :P From http://www.orafaq.com/wiki/Oracle_database_FAQ#How_do_I_find_the_overall_database_size.3F

  • Can we reduce the size of the disk having the Log files for a Dag Database

    There is an issue with disk space filling up for 4 databases part of the same DAG, each having 1 non lagged passive copy.
    The Disks containing the log files are from the VSphere Storage. The Disk size was temporarily expanded to avoid any outages.
    There is a full backup running currently, which is expected to clear the transaction logs on completion and that should be reducing the disk space utilized.
    The storage guys want to know whether they can reclaim the temporarily expanded disk size. i.e reduce the disk space from the storage containing Log Files without affecting anything.
    I couldn't find any documentation on this specific requirement, and want to confirm

    There is an issue with disk space filling up for 4 databases part of the same DAG, each having 1 non lagged passive copy.
    The Disks containing the log files are from the VSphere Storage. The Disk size was temporarily expanded to avoid any outages.
    There is a full backup running currently, which is expected to clear the transaction logs on completion and that should be reducing the disk space utilized.
    The storage guys want to know whether they can reclaim the temporarily expanded disk size. i.e reduce the disk space from the storage containing Log Files without affecting anything.
    I couldn't find any documentation on this specific requirement, and want to confirm
    I dont see why not. Once the logs are cleared, Exchange doesnt care.
    Twitter!: Please Note: My Posts are provided “AS IS” without warranty of any kind, either expressed or implied.

  • How to reduce the database size after deleting huge amount of rows?

    Hi,
    I have a large database. I removed almost half of the data/rows. Now i need to reduce the size of the database file as I need more disk space for the database file.
    What should I do in details please.
    Thanks.

    Hi,
    I have a large database. I removed almost half of the data/rows. Now i need to reduce the size of the database file as I need more disk space for the database file.
    What should I do in details please.
    Thanks.
    Deleting large data would have ultimately put ghost cleanup into action. You would not be able to free space using shrink operation untill it completes. So wait for time and then start shrinking. You should note that shrinking would cause fragmentation and
    you would have to rebuild indexes after shrinking is completed
    Please mark this reply as answer if it solved your issue or vote as helpful if it helped so that other forum members can benefit from it
    My Technet Wiki Article
    MVP

  • Want to reduce the size of database

    Hi Guys,
    We have Oracle 10G database over Unix platform, Customer want to reduce the size of database as much as possible and the ami of customer to move the storage area of this database to other one. so we resize some datafiles and get lots of free space at mount point but while checking the utilzation of table is showing somw what different as other . kindly the below O/P:
                                                                                        %                         MaxPoss    Max
    Tablespace Name                         KBytes            Used            Free   Used         Largest          Kbytes   Used
    *a DATA                             45,875,200       8,740,992      37,134,208   19.1       1,728,512     100,663,248   45.6
    *a HIGH_S_DATA                      21,504,000       1,331,520      20,172,480    6.2       3,048,704               0     .0
    *a HIGH_S_IND                       15,360,000         853,568      14,506,432    5.6       1,661,504               0     .0
    *a IND_DATA                         57,241,600       3,512,832      53,728,768    6.1       3,995,648     100,663,248   56.9
    *a LOW_S_DATA                      357,376,000      28,436,736     328,939,264    8.0       4,063,232     167,772,080  213.0
    *a LOW_S_IND                       268,902,400      12,148,352     256,754,048    4.5       4,063,232      67,108,832  400.7
    *a SYSAUX                            1,433,600         897,152         536,448   62.6         485,312      33,554,416    4.3
    *a UNPARSED_DATA                       102,400             128         102,272     .1         102,272               0     .0
    *a USERS                             4,096,000              64       4,095,936     .0       4,063,232      33,554,416   12.2
    *m SYSTEM                            1,433,600         835,520         598,080   58.3         597,952      33,554,416    4.3
    *m TEMP                             54,033,408      54,033,408               0  100.0               0      33,554,416  161.0
    *m UNDOTBS1                         37,650,416         426,416      37,224,000    1.1       4,063,232      33,554,416  112.2
    sum                                865,008,624     111,216,688     753,791,936
    SQL> select sum(bytes/1024/1024/1024) from dba_segments;
    SUM(BYTES/1024/1024/1024)
                   54.5392456
    SQL> select sum(bytes/1024/1024/1024) from dba_data_files;
    SUM(BYTES/1024/1024/1024)
                   773.406235above all o/p is different, no able to understand it. please help me in this .
    Could you please tell me is there any way to reset the HWM at Datafile level and how we reset the HWM of those tables having Materlized view?

    Hello,
    I'm not really sure what is the problem here? Everything seems to be OK. From what I have seen ,your database can stand to be resized down quite a bit.
    The actual amount it can be resized down depends on the highet HWM for the segments in the datafile, as in:
    set line 200col file_name format a90
    select file_name,
    ceil( (nvl(hwm,1)*&&blksize)/1024/1024 ) USED_SPACE,
    ceil( blocks*&&blksize/1024/1024) SIZE,
    ceil( blocks*&&blksize/1024/1024) -
    ceil( (nvl(hwm,1)*&&blksize)/1024/1024 ) ECONOMY --how much of the space can be reclaimed
    from dba_data_files a,
    ( select file_id, max(block_id+blocks-1) hwm -- here we get the max hwm for the datafile from the extents residing within it
    from dba_extents
    group by file_id ) b
    where a.file_id = b.file_id(+)
    and ceil( blocks*&&blksize/1024/1024) -
    ceil( (nvl(hwm,1)*&&blksize)/1024/1024 ) > 2000
    >
    You can also do a reorganization of the segment you suspect has a high degree of reclaimable space within its extents (high HWM and low actual space used). You can run segment advisor to help with that.
    Hope it helps,

  • I ahve a ipod touch of first generation model no.=MC540ll and i want to update my IOS of my ipod to 6.1.5 it requires 2 GB to start i want to know that after instillation the memory will be reduced or not?

    I ahve a ipod touch of first generation model no.=MC540ll and i want to update my IOS of my ipod to 6.1.5 it requires 2 GB to start i want to know that after instillation the memory will be reduced or not?

    A model MC540ll is a 4G iPod and can go to iOS 6.1.5.
    Updating via wifi (settings>General>Software Update requires about 2.4 GB of storage available. After the update all but maybe 50 MB will be returned to for use. iOS 6 takes up only a little more that iOS 5 but the excess storage space is needed to perform the update.
    If you update via iTunes yo do not need exces storage space

Maybe you are looking for

  • Problem ceating domains in 10.1.3.1.0

    I recently installed the OSA version of Oracle BPEL 10.1.3.1.0 everything was going well until I tried to create a new domain. I used the instructions in the online manual and created the domain through the http://localhost:port/BPELAdmim page. Every

  • Save an excel file by OLE2 object

    Hi everybody, I wanna save an excel file generated bye ole2 objects just same as the SaveAs dialog prompt. But SaveAs prompt is not working....

  • Styling Text submitted with CMS

    Hi I am designing a website with CMS for a local church. Could anyone help me with controlling the appearance of text submitted to the database using the record insert wizard. My primary problem is that any text submitted to the database does not dis

  • Problems with my New N9

    Hello, I have identified the following problem with my new N9 Map and drive icons are not opening I can not download from nokia store When it is not in use, the screen turns to green and drains the battery very fast. I need anti virus, Virus has atta

  • H.264 mov to Flash

    Adobe says Flash now "supports H.264 encoding". Since this encoding is QuickTime engine (ISO standard), how do I get an H.264-compressed MOV file to play in Flash Player? (e.g "import to Flash, export as FLV" or some such). Thanks in advance.