Memory usage growing up constantly

Oracle 8.1.7 EE on Windows 2000 Advanced Server. RAM 512, CPU Intel Xeon 2.4 Ghz.
The memory usage amount reported for the ORACLE.EXE at the Task Manager, in the Processes tab, in the Memory Usage column; is growing up constantly. Aproximately every week grow up too much and consume all the server memory resources. This cause the overall performance of the server goes down.
The "solution" is to stop and restart the OracleServiceSID service (and shutdown and restart up the instance), but sometimes is necessary to shutdown and restart the server.
Any idea about what is the problem and how to avoid this behavior?

Hi Ice,
The loaded swf file contains only a small photo, just for testing, with no listeners or other stuff in it.
Vic.

Similar Messages

  • Mail.app and AB Memory Usage Growing and Growing

    I've noticed this over the past week, but I don't know what could have triggered it. Both Address Book's and Mail's memory usage are growing out of control. Since I can easily not run AB much, it's not a big deal, but Mail.app is my preferred email application and I need to run it all the time.
    When I start Mail.app, it's around 65M RSIZE, but if I leave it running for a few hours, it quickly balloons up to 2+GB in RSIZE. I'm sure it'd keep getting bigger if I didn't restart it.
    This is with one account (Exchange) on a 6,2 MacBook Pro. I'm not currently using any plugins for Mail.app.
    Any idea what might be causing these problems, how I can help diagnose it, and how I might be able to fix them?
    Thanks,
    Pete

    Check out this conversation, there are some potential solutions:
    http://discussions.apple.com/thread.jspa?messageID=11446708&#11446708

  • Unloading external swf file keeps memory usage growing up

    Hello everybody,
    The idea is very simple: I have a submain.swf with 5 buttons. Each button loads an external .swf file.
    com_loader.load(URLRequest);
    After playing first external swf, before I go to next one, I try to unload it:
    1. remove all its associated listeners:
    com_loader.contentLoaderInfo.removeEventListener(Event.INIT-PROGRESS-COMPLETE, idFunction);
    2. unload it:
    com_loader.unload();
    3. set to null:
    com_loader = null;
    and load next external swf file and so on.
    The problem is that the FireFox/IE8.0 Memory Usage in Task Manager is going up and I see no sign of GC activity even after 5 mins of playing. Can anyone light me up pls? This is the second week working around and no good news until now.
    Thanks in advance.
    P.S.
    If I try com_loader.unloadAndStop(); it gives me the error: Error #2044: Unhandled IOErrorEvent:. text=Error #2036: Load Never Completed.
    Vince.

    Hi Ice,
    The loaded swf file contains only a small photo, just for testing, with no listeners or other stuff in it.
    Vic.

  • Network stream fxp excess memory usage and poor performance

    I'm trying to stream some datas à highspeed rate (3 channels à 1Mbytes/s) from my 9030 to my windows host. Because i don't need to use data on the rt side, i choose to forward FXP <+-,24,5> to my host throug a network stream.
    To avoid data loose i choose to use a wide buffer of 6000000 with this buffer my memory usage grow from 441mo to 672Mo and my rio is unable to stream the data. 
    With sgl or double, memory usage is 441 to 491Mo and datas can be streamed continusly.
    Anyone have encoutered this problem?

    SQL Developer is java based and relies on the jvm's memory management.
    I'm not aware of any memory leaks as such, but memory tends not to be returned to the system.
    Queries which return large return sets tend to use a lot of memory (SQL Developer has to build a java table containing all the results for display).
    You can restrict the maximum memory allocated by modifying settings in in <sqldeveloper>\ide\bin\ide.conf
    The defaults are -
    AddVMOption -Xmx640M
    AddVMOption -Xms128M

  • Problem with Firefox and very heavy memory usage

    For several releases now, Firefox has been particularly heavy on memory usage. With its most recent version, with a single browser instance and only one tab, Firefox consumes more memory that any other application running on my Windows PC. The memory footprint grows significantly as I open additional tabs, getting to almost 1GB when there are 7 or 8 tabs open. This is as true with no extensions or pluggins, and with the usual set, (firebug, fire cookie, fireshot, XMarks). Right now, with 2 tabs, the memory is at 217,128K and climbing, CPU is between 0.2 and 1.5%.
    I have read dozens of threads providing "helpful" suggestions, and tried any that seem reasonable. But like most others who experience Firebug's memory problems, none address the issue.
    Firefox is an excellent tool for web developers, and I rely on it heavily, but have now resorted to using Chrome as the default and only open Firefox when I must, in order to test or debug a page.
    Is there no hope of resolving this problem? So far, from responses to other similar threads, the response has been to deny any responsibility and blame extensions and/or pluggins. This is not helpful and not accurate. Will Firefox accept ownership for this problem and try to address it properly, or must we continue to suffer for your failings?

    55%, it's still 1.6Gb....there shouldn't be a problem scanning something that it says will take up 300Mb, then actually only takes up 70Mb.
    And not wrong, it obviously isn't releasing the memory when other applications need it because it doesn't, I have to close PS before it will release it. Yes, it probably is supposed to release it, but it isn't.
    Thank you for your answer (even if it did appear to me to be a bit rude/shouty, perhaps something more polite than "Wrong!" next time) but I'm sitting at my computer, and I can see what is using how much memory and when, you can't.

  • TOP - Pageouts and memory usage.

    Looks like I have plenty of memory for what I am doing. But What do the cumulative numbers around pagein's mean?
    Processes: 80 total, 7 running, 3 stuck, 70 sleeping... 523 threads 15:20:14
    Load Avg: 1.99, 1.53, 1.37 CPU usage: 10.33% user, 9.00% sys, 80.67% idle
    SharedLibs: num = 5, resident = 13M code, 0 data, 1468K linkedit.
    MemRegions: num = 32848, resident = 2116M + 18M private, 877M shared.
    PhysMem: 1853M wired, 2116M active, 3466M inactive, 7439M used, 753M free.
    VM: 19G + 284M 332461(0) pageins, 47(0) pageouts
    PID COMMAND %CPU TIME #TH #PRTS #MREGS RPRVT RSHRD RSIZE VSIZE
    12256 top 11.1% 0:14.05 1 18 40 684K 820K 1392K 18M
    12238 bash 0.0% 0:00.01 1 14 19 276K 184K 916K 18M
    12237 login 0.0% 0:00.02 1 17 136 908K 8724K 3164K 32M
    12236 Terminal 1.7% 0:01.45 3 99 433 5264K+ 49M+ 21M+ 419M+
    12225 SyncServer 0.0% 0:00.15 2 50 290 3512K 41M 12M 88M
    11248 WinAppHelp 0.0% 0:00.40 1 51 289 3208K 40M 11M 339M
    11027 Skype 0.0% 0:19.47 31 313 865 46M 76M 88M 557M
    10315 PrinterPro 0.0% 0:00.83 2 85 476 12M 51M 29M 423M
    10313 launchd 0.0% 0:01.25 3 24 25 120K 296K 464K 18M
    8626 QuickTime 0.0% 17:21.09 5 187 655 13M 67M 43M 486M
    8556 DashboardC 0.0% 0:06.31 4 124 538 16M 58M 38M 436M
    8555 DashboardC 0.0% 0:10.95 4 149 545 23M 58M 49M 444M
    8534 Livestatio 16.8% 59:00.13 12 198 771 30M- 94M 62M 584M
    8220 Mail 0.0% 0:43.91 27 323 968 24M 94M 80M 554M
    2806 Python 0.8% 5:20.05 2 33 281 13M- 248K 14M- 36M-
    2761 Transmissi 0.9% 2:49.17 21 270 773 20M 90M 56M 506M

    About OS X Memory Management and Usage
    Reading system memory usage in Activity Monitor
    Memory Management in Mac OS X
    Performance Guidelines- Memory Management in Mac OS X
    A detailed look at memory usage in OS X
    Understanding top output in the Terminal
    The amount of available RAM for applications is the sum of Free RAM and Inactive RAM. This will change as applications are opened and closed or change from active to inactive status. The Swap figure represents an estimate of the total amount of swap space required for VM if used, but does not necessarily indicate the actual size of the existing swap file. If you are really in need of more RAM that would be indicated by how frequently the system uses VM. If you open the Terminal and run the top command at the prompt you will find information reported on Pageins () and Pageouts (). Pageouts () is the important figure. If the value in the parentheses is 0 (zero) then OS X is not making instantaneous use of VM which means you have adequate physical RAM for the system with the applications you have loaded. If the figure in parentheses is running positive and your hard drive is constantly being used (thrashing) then you need more physical RAM.

  • Memory problems: computer freezes on heavy memory usage

    Ever since I changed the internal hard drive of my MacBook Pro (I installed a Seagate Barracuda 1Tb disk), I have the following problem:
    Whenever an application needs a lot of memory (for example Parallels installing a new Windows system, or Acrobat Pro optimizing a big PDF file), the computer progressively freezes. What I mean by freezing is that I can still move the mouse cursor, but clicking has no effect, and there is no interaction whatsoever with the GUI. Even the clock on the menu bar stops.
    I never had the problem before. I guess that under normal conditions the Finder will always keep a few Mb of RAM for its own use. After all I have 4Gb of RAM installed.
    Is there some way I can prevent the computer from freezing? Has this something to do with VM? Ever since I changed the hard disk, this problem happens once or twice every day, and the only solution is to force a restart, losing all my work…
    Thanks in advance for any advice…

    First, do not assign more than half your available RAM to a VM. It's best to let the VM software's configuration to select the amount used by its defaults.
    Second, watch how many concurrent applications you attempt to use. Monitor your memory usage in Activity Monitor. If the Available RAM (Inactive RAM plus Free RAM) is low and Free RAM is near zero, then you have too many concurrent applications running. This forces the system to start using the disk-based virtual memory file. With too many apps swapping into the vm file disk thrashing will occur that ultimately can lead to the computer appearing to freeze up or to actually freeze up.
    Of course your alternative is to install more RAM if in fact you are over-taxing what you have now. You may find the following helpful reading:
    About OS X Memory Management and Usage
    Reading system memory usage in Activity Monitor
    Memory Management in Mac OS X
    Performance Guidelines- Memory Management in Mac OS X
    A detailed look at memory usage in OS X
    Understanding top output in the Terminal
    The amount of available RAM for applications is the sum of Free RAM and Inactive RAM. This will change as applications are opened and closed or change from active to inactive status. The Swap figure represents an estimate of the total amount of swap space required for VM if used, but does not necessarily indicate the actual size of the existing swap file. If you are really in need of more RAM that would be indicated by how frequently the system uses VM. If you open the Terminal and run the top command at the prompt you will find information reported on Pageins () and Pageouts (). Pageouts () is the important figure. If the value in the parentheses is 0 (zero) then OS X is not making instantaneous use of VM which means you have adequate physical RAM for the system with the applications you have loaded. If the figure in parentheses is running positive and your hard drive is constantly being used (thrashing) then you need more physical RAM.

  • Massive memory usage when using latest version of final cut, 10 minutes after restart.

    Hey I am really struggling with using Final Cut Pro X and I suspect there is some software issue.
    I am editing a one minute video clip, the video files are 1080 h264. I restart my computer and only open final cut. After 10 minutes of editing I cant barely navigate my computer anymore as it is going so slow, until eventually I have to force reboot my computer. I can scarcly export anything out of final cut, when I do my computer feels like its about to die and if it exports I seem to get broken frames and/or missing media.
    I reinstalled my mac only about 2 months ago.
    System specs.
    Mac book pro.
    13-inch, Late 2011
    Processor  2.4 GHz Intel Core i5
    Memory  4 GB 1333 MHz DDR3
    Graphics  Intel HD Graphics 3000 384 MB
    On my previous mac book pro (early 2010, 4 gm ram), running snow leopard and previous versions of final cut I was able to edit 3, 1 hour documentaries with very little problem, but now I cant even manage a one minute clip.
    Not sure what to do. Here is a screen shot of my memory usage. Im getting massive swap memory and lots of inactive memory. It seems very strange. Especially after only 10 minutes running only final cut.

    FCP is only using 409 MBs of memory out of your 4 GBs. But whatever else you have open is using up virtually all the physical RAM you have. This is forcing the memory manager to page memory out to the disk-based VM file. That's a major reason for major slowdowns.
    You have 4 GBs of RAM. How many applications are you running concurrently? If you look at Page outs: you will note the positive number in parentheses. That means your computer is hitting the hard drive a lot because the memory demands are too high. If possible cut down on how many concurrent applications you run or put more RAM into the computer, if that's possible.
    About OS X Memory Management and Usage
    Using Activity Monitor to read System Memory & determine how much RAM is used
    Memory Management in Mac OS X
    Performance Guidelines- Memory Management in Mac OS X
    A detailed look at memory usage in OS X
    Understanding top output in the Terminal
    The amount of available RAM for applications is the sum of Free RAM and Inactive RAM. This will change as applications are opened and closed or change from active to inactive status. The Swap figure represents an estimate of the total amount of swap space required for VM if used, but does not necessarily indicate the actual size of the existing swap file. If you are really in need of more RAM that would be indicated by how frequently the system uses VM. If you open the Terminal and run the top command at the prompt you will find information reported on Pageins () and Pageouts (). Pageouts () is the important figure. If the value in the parentheses is 0 (zero) then OS X is not making instantaneous use of VM which means you have adequate physical RAM for the system with the applications you have loaded. If the figure in parentheses is running positive and your hard drive is constantly being used (thrashing) then you need more physical RAM.
    Adding RAM only makes it possible to run more programs concurrently.  It doesn't speed up the computer nor make games run faster.  What it can do is prevent the system from having to use disk-based VM when it runs out of RAM because you are trying to run too many applications concurrently or using applications that are extremely RAM dependent.  It will improve the performance of applications that run mostly in RAM or when loading programs.

  • High memory usage on JDBC 10.2.0.1.0 driver on Prepared/Callable Statements

    We are observing high memory usage for each callable/prepared stmt, using 10.2.0.1.0 JDBC Driver. The char[] in oracle/jdbc/driver/T4CVarcharAccessor was alloted 64K to 320K and grows with usage. This problem is worse with 10.1.0.2. driver which was alloted 720K byte of memory for each stmt right at the start.
    We found this by doing a JVM heap dump and analyzing the heap dump using IBM's heap analyser. Here is a snapshot of the heap dump for this object:
    321,240 [216] 11 oracle/jdbc/driver/T4CVarcharAccessor 0x72752968
    - 320,616 [320,616] 0 char[] 0x72761028
    - 216 [216] 0 short[] 0x727527d8
    - 72 [32] 1 java/lang/String 0x727530a0
    - 24 [24] 0 int[] 0x72752938
    - 24 [24] 0 int[] 0x72752948
    - 24 [24] 0 int[] 0x72752958
    - 16 [16] 0 bool[] 0x72752928
    - 16 [16] 0 byte[] 0x727528b0
    - 16 [16] 0 bool[] 0x72752918
    - 10,336 [88] 15 oracle/jdbc/driver/T4CMAREngine 0x712e7128
    - 1,544 [1,032] 79 oracle/jdbc/driver/T4CPreparedStatement 0x72754c58
    It is repeated many times for each prepared/callable stmt call.
    Details of our platform is:
    Database - Oracle Database 10g Release 10.2.0.1.0 - 64bit Production
    JDBC Driver - Oracle Database 10g Release 2 (10.2.0.1.0) JDBC Drivers
    JDK - [Classic VM, Version 1.4.2] from [IBM Corporation]
    Our callable stmts are not using any of the Oracle caching facility. It is a simple call stmt with OUT parameters and the stmt is closed after each execution. However, we implement our own connection pooling and do not close the connection after each stmt.
    Is there a workaround to this? Would appreciate any feedback.

    What is happening is that each new CallableStatement you create allocates a new char[]. I would strongly encourage you to use the implicit statement cache if at all possible. That way instead of creating a new statement each time with a new char[] you will get an already existing statement and reuse the existing char[]. Closing a statement releases the char[] so if you really are closing the statements the char[]s should be GC'd.
    Douglas

  • High Eden Java Memory Usage/Garbage Collection

    Hi,
    I am trying to make sure that my Coldfusion Server is optomised to the max and to find out what is normal limits.
    Basically it looks like at times my servers can run slow but it is possible that this is caused by a very old bloated code base.
    Jrun can sometimes have very high CPU usage so I purchased Fusion Reactor to see what is going on under the hood.
    Here are my current Java settings (running v6u24):
    java.args=-server -Xmx4096m -Xms4096m -XX:MaxPermSize=256m -XX:PermSize=256m -Dsun.rmi.dgc.client.gcInterval=600000 -Dsun.rmi.dgc.server.gcInterval=600000 -Dsun.io.useCanonCaches=false -XX:+UseParallelGC -Xbatch ........
    With regards Memory, the only memory that seems to be running a lot of Garbage Collection is the Eden Memory Space. It climbs to nearly 1.2GB in total just under every minute at which time it looks like GC kicks in and the usage drops to about 100MB.
    Survivor memory grows to about 80-100MB over the space of 10 minutes but drops to 0 after the scheduled full GC runs. Old Gen memory fluctuates between 225MB and 350MB with small steps (~50MB) up or down when full GC runs every 10 minutes.
    I had the heap set to 2GB initally in total giving about 600MB to the Eden Space. When I looked at the graphs from Fusion Reactor I could see that there was (minor) Garbage Collection about 2-3 times a minute when the memory usage maxed out the entire 600MB which seemed a high frequency to my untrained eye. I then upped the memory to 4GB in total (~1.2GB auto given to Eden space) to see the difference and saw that GC happened 1-2 times per minute.
    Is it normal in Coldfusion that the Eden memory would grow so quickly and have garbage collection run so often? i.e do these graphs look normal?
    Also should I somehow redistribute the memory available to give the Eden memory more since it seems to be where all the action is?
    Any other advice for performance improvements would be much appreciated.
    Note: These graphs are not from a period where jrun had high CPU.
    Here are the graphs:
    PS Eden Space Graph
    PS Survivor Space Graph
    PS Old Gen Graph
    PS Perm Gen Graph
    Heap Memory Graph
    Heap/Non Heap Memory Graph
    CPU Graph
    Request Average Execution Time Graph
    Request Activity Graph
    Code Cache Graph

    Hi,
    >Is it normal in Coldfusion that the Eden memory would grow so quickly and have garbage collection run so often?
    Yes normal to garbage collect Eden often. That is a minor garbage collection.
    >Also should I somehow redistribute the memory available to give the Eden memory more since it seems to be where all the action is?
    Sometimes it is good to set Eden (Eden and its two Survivor Spaces combined make up New or Young Generation part of JVM heap) to a smaller size. I know your thinking - what make it less, but I want to make it bigger. Give less a try (sometimes less = more, bigger not = better) and monitor the situation. I like to use -Xmn switch, some sources say to use other method/s. Perhaps you could try java.args=-server -Xmx4096m -Xms4096m -Xmn172m etc. I better mention make a backup copy of jvm.config before applying changes. Having said that now you know how you can set the size to bigger if you want.
    I think the JVM is perhaps making some poor decisions with sizing the heap. With Eden growing to 1Gb then being evacuated not many objects are surviving and therefore not being promoted to Old Generation. This ultimately means the object will need to be loaded again latter to Eden rather than being referenced in the Old generation part of the heap. Adds up to poor performance.
    >Any other advice for performance improvements would be much appreciated.
    You are using Parallel garbage collector. Perhaps you could enable that to run multi-threaded reducing the time duration of the garbage collections, jvm args ...-XX:+UseParallelGC -XX:ParallelGCThreads=N etc where N = CPU cores (eg quad core = 4).
    HTH, Carl.

  • How to minimize memory usage with VM settings

    Using JDK 1.4.2, I am trying to minimize the memory usage of my application.
    Our GUI application consumes about ~20MB of heap space during normal operation, but it can easily grow to 50MB in some rare cases (Special feature accessed by user).
    We have to be able to run multiple instances (separate process/VMs) of this application on low end Windows PCs.
    The applications don't usually run for very long (less than 8 hours).
    Here is my problem.
    If I set the -Xms=20M -Xmx=64M, it works fine, but the GC is not very aggressive and soon, the application starts using a lot of memory, even if it is not really needed. And it never returns the memory to Windows OS.
    So each VM take more memory than really needed, which causes significant swapping on these low end PC (little RAM available).
    If I reduce the -Xmx flag, than I run out of memory in the rare cases where I need the full memory.
    Ideally, I would like to configure the GC to be more aggressive, i.e. the GC should collect instead of getting additional memory from the OS when possible.
    I read a few papers on Sun's web site about HotSpot tunning, but most of them are geared for large servers.
    Thanks!
    Francois

    And you are using the task manager to determine how "much" memory is being used correct?
    I suggest that you search out and read the numerous threads on this site that discuss that subject.

  • SQL Server Memory Usage Peaks to 95% and its not releasing SQL 2012

    We are currently running SQL 2012  64 bit, Lock Pages Enabled, 128 GB. We allocated Max Memory 112000, Min Memory to 0. We have a SQL Maintenace Job which backs up (Full Backup) of our server around 11PM at that time the SQL Server memory peaks to almost
    96% even during full business hours we are constantly in the range of 56 to 60% but after 11PM within 10 minutes of the job it peaks to 95 or 96 Percentage and even after the job completes it's not releasing the memory back. I have to manually shut the sql
    service and then restart at that time it comes back to normal.
    Any suggestions/any help really appreciated.

    Here are the details Memory is at 96% I still didn't restart...Please any help appreciated...The memory detals is for 24 hours  I restarted the service around 11AM before I posted my previous message..
    Memory Used by SqLServer: 119329
    Locked Pages Used SQLServer : 105969
    Total VAS in MB : 8388607
    Process Physical Memory Low : 0
    Process Virtual Memory Low : 0
    Max Memory set to :112000
    Out of 128 Memory Last 24 hours the memory usage in %
    Percent Memory Used - Total Memory is 128.0 GB
    Last 24 Hours
    DATE / TIME
    Memory Used
    19-Dec-14
    11:46 PM
    96.24659602
    20-Dec-14
    12:46 AM
    96.24578349
    20-Dec-14
    1:46 AM
    96.25146739
    20-Dec-14
    2:46 AM
    96.24345652
    20-Dec-14
    3:46 AM
    96.27311834
    20-Dec-14
    4:46 AM
    96.28947067
    20-Dec-14
    5:46 AM
    96.18931325
    20-Dec-14
    6:46 AM
    96.09323502
    20-Dec-14
    7:46 AM
    96.07915497
    20-Dec-14
    8:46 AM
    96.07906977
    20-Dec-14
    9:46 AM
    96.0784111
    20-Dec-14
    10:46 AM
    96.07415009
    20-Dec-14
    11:46 AM
    26.03422141
    20-Dec-14
    12:46 PM
    33.57474359
    20-Dec-14
    1:46 PM
    39.466561
    20-Dec-14
    2:46 PM
    41.85940742
    20-Dec-14
    3:46 PM
    43.89071274
    20-Dec-14
    4:46 PM
    45.80877368
    20-Dec-14
    5:46 PM
    46.49493281
    20-Dec-14
    6:46 PM
    46.68486468
    20-Dec-14
    7:46 PM
    46.69701958
    20-Dec-14
    8:46 PM
    46.69994736
    20-Dec-14
    9:46 PM
    57.5012455
    20-Dec-14
    10:46 PM
    96.25695419
    I verified its a sqL job and still my memory is 95%
    It peaeked at 10:46PM and here are the details of the SQL job which started at 10:30 
    Progress: 2014-12-20 22:30:04.39 
          Source: Check Database Integrity Task      Executing query "USE [DATASTORE]  ".: 50% complete  End Progress  Progress: 2014-12-20 22:43:06.10 
          Source: Check Database Integrity Task      Executing query "DBCC CHECKDB(N'DATASTORE')  WITH NO_INFOMSGS  ".: 100% complete  End Progress
        Progress: 2014-12-20 22:43:06.11     Source: Check Database Integrity Task      Executing query "USE [ETL_Proc]  ".: 50% complete 
     End Progress  Progress: 2014-12-20 22:46:52.56     Source: Check Database Integrity Task      Executing query "DBCC CHECKDB(N'ETL_Proc') 
      WITH NO_INFOMSGS  ".: 100% complete  End Progress  Progress: 2014-12-20 22:46:52.64     Source: Back Up Database Task
            Executing query "EXECUTE master.dbo.xp_create_subdir N'P:\SQL_Backu...".: 20% complete  End Progress  
    Progress: 2014-12-20 22:46:52.64     Source: Back Up Database Task      Executing query "EXECUTE master.dbo.xp_create_subdir N'P:\SQL_Backu...".
    : 40% complete  End Progress  Progress: 2014-12-20 22:46:52.64   
      Source: Back Up Database Task      Executing query "EXECUTE master.dbo.xp_create_subdir N'P:\SQL_Backu...".: 60% complete 
       End Progress  Progress: 2014-12-20 22:46:52.64     Source: Back Up Database Task    
     Executing query "EXECUTE master.dbo.xp_create_subdir N'P:\SQL_Backu...".: 80% complete  End Progress  
     Progress: 2014-12-20 22:46:52.64     Source: Back Up Database Task      
     Executing query "EXECUTE master.dbo.xp_create_subdir N'P:\SQL_Backu...".: 100% complete  End Progress 
      Progress: 2014-12-20 22:46:55.63     Source: Back Up Database Task    
        Executing query "BACKUP DATABASE [ReportServer] TO  DISK = N'P:\SQL...".: 100% complete  
    End Progress  Progress: 2014-12-20 22:46:56.55     Source: Back Up Database Task    
      Executing query "BACKUP DATABASE [ReportServerTempDB] TO  DISK = N'...".: 100% complete  End Progress  Progress: 2014-12-20 22:46:57.35  
         Source: Back Up Database Task      Executing query "BACKUP DATABASE [dbamaint] TO  DISK = N'P:\SQL_Bac...".: 100% complete  End Progress 
      Progress: 2014-12-20 22:51:13.08     Source: Back Up Database Task   
         Executing query "BACKUP DATABASE [DATASTORE] TO  DISK = N'P:\SQL_Ba...".: 100% complete  End Progress
       Progress: 2014-12-20 22:51:52.72     Source: Back Up Database Task     
    Executing query "BACKUP DATABASE [ETL_Proc] TO  DISK = N'P:\SQL_Bac...".: 100% complete  End Progress  Progress: 2014-12-20 22:51:54.87 
        Source: Rebuild Index Task      Executing query "USE [ReportServer]  ".: 0% complete  End Progress  Progress:
     2014-12-20 22:51:54.88     Source: Rebuild Index Task      Executing query "ALT...  The package executed successf...  The step succeeded.

  • SQL Memory Usage

    Hello
    I am updating some items with the DTW. i am updating 9000 items. I have made 9 files of 1000 records, i am just updating the picturname field. When i start to upload the files the SQL Memory usage starts to grow more and more. When i try to upload the third file the memory has grown too much, and the DTW says the file update ok but it doesnt update the database. So i need to restart the SQL service (mssqlserver) and do it again. Some times it doesnt work so i need to restart the server.
    Any ideas??
    Jacobo

    hi,
    Check this information abt DI API Memory Consumption in WIKI SDK FAQ's.
    https://www.sdn.sap.com/irj/scn/wiki?path=/display/b1/faq_sdk
    DI API
    Memory consumption
    New connection mechanism in 2007 vs. classical
    1. In the old method the DI API was loaded into the same process with the Add-On and was an "actual part" of it, so calls to the DI-API were very quick and direct
    2. In the new method there is one common DI-API which is part of the Core B1. All the Add-Ons will use the same DI-API IF (!!!) they work in the new method. The calls to the DI-API have now to "go between" two processes. This means that they go through more processing and although it's all in the same machine and no actual communication (i.e., network traffic) is actually happening, the system's CPU and memory are "working harder".
    The impact is extremely small on an individual call level, but for an Add-On that makes a large amount of calls this difference accumulates...
    There is no huge additional CPU or Memory consumption. Most of the impact is on the Response Time level. Some of it is CPU consumption and some of it is Context Switch waiting.
    3. This trade-off between Memory consumption and Response Time is actually another reason why R&D thought it's a good idea to leave the new method to be optional based on a decision from the developer.
    Jeyakanthan

  • Solaris process memory usage increase but not forever

    On Solaris 10 I have a multithreaded process with a strange behaviour. It manages complicated C++ structures (RWTVal or RWPtr). These structures are built from data stored in a database (using Pro*C). Each hour the process looks for new informacion in database, builds new structures on memory and it frees older data. But, each time it repeats this procedure, the process memory usage increases several MB (12/16MB). Process's memory usage starts from 100M until near 1,4G. Just to this point, it seems the process has memory leaks. But the strange behaviour is that after this point, the process stops to continue growing up anymore. When I try to look for memory leaks (using Purify tool) the process doesn't grow up and no significant leaks were showed. Did anyone found a similar behaviour or can explain what could be happening?

    markza wrote:
    Hi, thanks for responding
    Ja, i guess thats possible, but to do it all one row by row seems ridiculous, and it'll be so time consuming and sluggish surely.  I mean, for a months worth of data (which is realistic) thats 44640 individual queries.  If push comes to shove, then I'll have to try that for sure.  
    You can see by the example that I'm saving it to a text file, in csv format.  So it needs to be a string array, a cluster won't be of much help I dont think.
    The only other way I can think of is to break it up into more manageable chunks...maybe pull each column separately in a for loop and build up a 2D array like that until the spreadsheet storing.  
    You only do 1 query, but instead of Fetching All (as the Select does) you'll use the cursor to step through the data.
    You can use Format to String or Write Spreadsheet fire with doubles.
    You can break it down to get the data day by day instead of a full month at once.
    /Y
    LabVIEW 8.2 - 2014
    "Only dead fish swim downstream" - "My life for Kudos!" - "Dumb people repeat old mistakes - smart ones create new ones."
    G# - Free award winning reference based OOP for LV

  • Memory Usage in WinTaskManger vs heap size

    hello,
    I have exactly the same problem as in the topic "Windows Task Manager vs. Heap", posted at Dec 10, 2004 5:17 AM, by jorgeHX.
    Here are the symptoms:
    1) My application starts at 20MB seen by windows task manager
    2) I use profiler to monitor the heap. Heap is always working very healthly - heap size in the profiler increases by a minimum until the gc comes around so that the used heap size drops down again.
    3) However, after doing a relatively memory-consuming operation (loop of String indexing, patterning ..etc.), the memory usage in windows task manager goes up couple of MBs but never drops down.
    4) Then I go manually free the heap (System.gc()), I can see GC is freeing heap. However, the memory in windows task manager remains no change, no matter how many times I force the garbabge collection.
    This is a bad thing - if my application keeps doing that memory-consuming operation again and again, the memory seen in the windows task manager will be growing and growing to hundreds of MBs until the windows alerts "low on virtual memory".
    I tried everything already. I set every instance = null at the end, I delete every reference but the memory just keeps increasing! WHY?????
    Anyone can help me charactorize the problem? I am so in dark!!
    Ryan-Chi

    I guess my problem can also be interpretated as
    "Why doesn't JVM return memory to OS?"
    It does, depending on the setting of the -XX:MaxHeapFreeRatio option. "Normal" operation using the default setting does not usually cause memory to be returned to the os.
    Search for this option term for explanations.

Maybe you are looking for

  • Error while deploying, Custom Style Skin in OBIEE 11.1.1.6.7

    Hi, I have deployed Custom Style Skin in OBIEE 11.1.1.5 successfuly (with the help of http://www.rittmanmead.com/2010/12/oracle-bi-ee-11g-styles-skins-custom-xml-messages/ ). Now when we're moving RPD, Catalog & Custom Style Skin from OBIEE 11.1.1.5

  • How do i delete old apple id's

    How do I delete old apple id's?  I have a new Iphone 5 and every time I try to download a new app my old apple id is presented and asks for my old password which I can't remember.  So frustrating.  I have a new apple id that works so I need to delete

  • Error in post goods receipt

    Hi...... There is an error " Update Control of Movement Type is Incorrect (entry 161XX_L_V) when i create Post Goods Receipt. Need help Thanks

  • Recommendations for a UPS?

    Can anyone recommend an inexpensive UPS that would be suitable for the power load of a Mac Pro with 4 hard drives? I was told by a salesman that I needed at least a 1500 watt UPS to support a Mac Pro, and the prices were about £200. They suggested an

  • Stocks App not saving changes

    My stocks app is not saving changes no matter what changes I make.  After making a change, I'll exit the app, re-enter and it defaults back to a whole bunch of stuff I didn't add. Can someone help me with this, I'm not sure where to even start!