Heap dump file - Generate to a different folder

Hello,
When the AS Java iis generating the heap dump file, is it possible to generate it to a different folder rather than the standard one: /usr/sap// ?
Best regards,
Gonçalo  Mouro Vaz

Hello Gonçalo
I don't think this is possible.
As per SAP Note 1004255;
On the first occurrence (only) of an OutOfMemoryError the JVM
will write a heap dump in the
/usr/sap/ directory
Can i ask why you would like it in a different folder?
Is it a space issue?
Thanks
Kenny

Similar Messages

  • Heap Dump file generation problem

    Hi,
    I've configured configtool to have these 2 parameters:
    -XX:+HeapDumpOnOutOfMemoryError
    -XX:+HeapDumpOnCtrlBreak
    In my understanding, with these 2 parameters, the heap dump files will only be generated under 2 situations, ie, when out of memory occurred, or user manually click CLTR + BREAK in MMC.
    1) Unfortunately, there are many heap dump files (9 in total) generated when none of the above situation occured. I couldnt find "OutOfMemoryError" in the default trace, nor the shallow heap size of those heap dump files are anywhere near the memory limit. The consequences are our server run out of disk space.
    My question is, what are the other possibilities that heap dump file will be generated?
    2) In the Memory Consumption graph (NWA (http://host:port/nwa) -> System Management -> Monitoring -> Java Systems Reports), out of memory error occurred when the memory usage is reaching about 80% of the allocated memory. What are the remaining 20% or so reserved for ?
    Any help would be much appreciated.
    Thanks.

    Hi,
    Having the -XX:+HeapDumpOnCtrlBreak option makes the VM trigger a heap dump, whenever a CTRL_BREAK event appears. The same event is used also to trigger a thread dump, an action you can do manually from the SAP Management Console, I think it is called "Dump stacks". So if there was someone triggering thread dumps for analysis of other types of problems, this has the side effect of writing also a heap dump.
    Additionally, the server itself may trigger a thread dump (and by this also a heap dump if the option is present). It does this for example when a timeout appears during the start or stop of the server. A thread dump from such a moment allows us to see for example which service is unable to start.
    Therefore, I would recommend that you leave only the -XX:+HeapDumpOnOutOfMemoryError, as long as you don't plan to trigger any heap dumps on your own. The latter will cause the VM to write a heap dump only once - on the first appearance of an OutOfMemoryError.
    In case you need to trigger the heap dumps manually, leave the -XX:+HeapDumpOnCtrlBreak option for the moment of troubleshooting, but consider if you want to keep it afterwards.
    If heap dumps were written because of an OutOfMemoryError you should be able to see this in the dev_server file in /usr/sap/<SID>/<inst>/work/ . Also there you should be able to see if indeed thread dumps were triggered (just search for "Full Thread ").
    I hope this helps.
    Regards,
    Krum

  • JVMDG315: JVM Requesting Heap dump file

    Hi expert,
    Using 10.2.04 Db with 11.5.10.2 on AIX 64 bit.
    While running audit report which generate pdf format output gone in error....
    JVMDG217: Dump Handler is Processing OutOfMemory - Please Wait.
    JVMDG315: JVM Requesting Heap dump file
    ..............................JVMDG318: Heap dump file written to /d04_testdbx/appltop/testcomn/admin/log/TEST_tajorn3/heapdump7545250.1301289300.phd
    JVMDG303: JVM Requesting Java core file
    JVMDG304: Java core file written to /d04_testdbx/appltop/testcomn/admin/log/TEST_tajorn3/javacore7545250.1301289344.txt
    JVMDG274: Dump Handler has Processed OutOfMemory.
    JVMST109: Insufficient space in Javaheap to satisfy allocation request
    Exception in thread "main" java.lang.OutOfMemoryError
         at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java(Compiled Code))
         at java.io.OutputStream.write(OutputStream.java(Compiled Code))
         at oracle.apps.xdo.common.log.LogOutputStream.write(LogOutputStream.java(Compiled Code))
         at oracle.apps.xdo.generator.pdf.PDFStream.output(PDFStream.java(Compiled Code))
         at oracle.apps.xdo.generator.pdf.PDFGenerator$ObjectManager.write(PDFGenerator.java(Compiled Code))
         at oracle.apps.xdo.generator.pdf.PDFGenerator$ObjectManager.writeWritable(PDFGenerator.java(Inlined Compiled Code))
         at oracle.apps.xdo.generator.pdf.PDFGenerator.closePage(PDFGenerator.java(Compiled Code))
         at oracle.apps.xdo.generator.pdf.PDFGenerator.newPage(PDFGenerator.java(Compiled Code))
         at oracle.apps.xdo.generator.ProxyGenerator.genNewPageWithSize(ProxyGenerator.java(Compiled Code))
         at oracle.apps.xdo.generator.ProxyGenerator.processCommand(ProxyGenerator.java(Compiled Code))
         at oracle.apps.xdo.generator.ProxyGenerator.processCommandsFromDataStream(ProxyGenerator.java(Compiled Code))
         at oracle.apps.xdo.generator.ProxyGenerator.close(ProxyGenerator.java:1230)
         at oracle.apps.xdo.template.fo.FOProcessingEngine.process(FOProcessingEngine.java:336)
         at oracle.apps.xdo.template.FOProcessor.generate(FOProcessor.java:1043)
         at oracle.apps.xdo.oa.schema.server.TemplateHelper.runProcessTemplate(TemplateHelper.java:5888)
         at oracle.apps.xdo.oa.schema.server.TemplateHelper.processTemplate(TemplateHelper.java:3593)
         at oracle.apps.pay.pdfgen.server.PayPDFGen.processForm(PayPDFGen.java:243)
         at oracle.apps.pay.pdfgen.server.PayPDFGen.runProgram(PayPDFGen.java:795)
         at oracle.apps.fnd.cp.request.Run.main(Run.java:161)
    oracle.apps.pay.pdfgen.server.PayPDFGen
    Program exited with status 1
    No any error in db alert ...
    Pleaes suggest where is prob..
    Thanks in ADV

    Hello,
    * JVMST109: Insufficient space in Javaheap to satisfy allocation request
    Exception in thread "main" java.lang.OutOfMemoryError
    This is your problem. There is no enough memory. Change Java heap settings
    Regards,
    Shomoos
    Edited by: shomoos.aldujaily on Mar 29, 2011 12:39 AM

  • Analyse large heap dump file

    Hi,
    I have to analyse large heap dump file (3.6GB) from production environment. However if open it in eclipse mat, it is giving OutOfMemoryError. I tried to increase eclipse workbench java heap size as well. But it doesnt help. I also tried with visualVM as well. Can we split the heap dump file into small size? Or is there any way to set max heap dump file size for jvm options so that we collect reasonable size of heap dumps.
    Thanks,
    Prasad

    Hi, Prasad
    Have you tried open in 64-bit mat on a 64-bit platform with large heap size and CMS gc policy in MemoryAnalyzer.ini file ? the mat is good toolkit on analysing java heapdump file, if it cann't works, you can try Memory Dump Diagnostic for Java(MDD4J) in 64-bit IBM Support Assistant with large heap size.

  • Heap dump file size vs heap size

    Hi,
    I'd like to clarify my doubts.
    At the moment we're analyzing Sun JVM heap dumps from Solaris platform.
    Observation is that heap dump file is around 1,1GB while after loading to SAP Memory Analyzer it displays statistics: "Heap: 193,656,968" which as I understood is size of heap.
    After I run:
    jmap -heap <PID>
    I get following information:
    using thread-local object allocation
    Parallel GC with 8 thread(s)
    Heap Configuration:
       MinHeapFreeRatio = 40
       MaxHeapFreeRatio = 70
       MaxHeapSize      = 3221225472 (3072.0MB)
       NewSize          = 2228224 (2.125MB)
       MaxNewSize       = 4294901760 (4095.9375MB)
       OldSize          = 1441792 (1.375MB)
       NewRatio         = 2
       SurvivorRatio    = 32
       PermSize         = 16777216 (16.0MB)
       MaxPermSize      = 67108864 (64.0MB)
    Heap Usage:
    PS Young Generation
    Eden Space:
       capacity = 288620544 (275.25MB)
       used     = 26593352 (25.36139678955078MB)
       free     = 262027192 (249.88860321044922MB)
       9.213949787302736% used
    From Space:
       capacity = 2555904 (2.4375MB)
       used     = 467176 (0.44553375244140625MB)
       free     = 2088728 (1.9919662475585938MB)
       18.27830779246795% used
    To Space:
       capacity = 2490368 (2.375MB)
       used     = 0 (0.0MB)
       free     = 2490368 (2.375MB)
       0.0% used
    PS Old Generation
       capacity = 1568669696 (1496.0MB)
       used     = 1101274224 (1050.2569427490234MB)
       free     = 467395472 (445.74305725097656MB)
       70.20434109284916% used
    PS Perm Generation
       capacity = 67108864 (64.0MB)
       used     = 40103200 (38.245391845703125MB)
       free     = 27005664 (25.754608154296875MB)
       59.75842475891113% used
    So I'm just wondering what is this "Heap" in Statistic Information field visible in SAP Memory Analyzer.
    When I go to Dominator Tree view, I look at Retained Heap column and I see that they roughly sum up to 193,656,968.
    Could someone put some more light on it?
    thanks
    Michal

    Hi Michal,
    that looks indeed very odd. First, let me ask which version do you use? We had a problem in the past where classes loaded by the system class loader were not marked as garbage collection roots and hence were removed. This problem is fixed in the current version (1.1). If it is version 1.1, then I would love to have a look at the heap dump and find out if it is us.
    Having said that, this is what we do: After parsing the heap dump, we remove objects which are not reachable from garbage collection roots. This is necessary, because the heap dump can contain garbage. For example, the mark-sweep-compact of the old/perm generation leaves some dead space in the form of int arrays or java.lang.Object to win time during the compacting phase: by leaving behind dead objects, not every live object has to be moved which means not every object needs a new address. This is the kind of garbage we remove.
    Of course, we do not remove objects kept alive only by weak or soft references. To see what memory is kept alive only through weak or soft references, one can run the "Soft Reference Statistics" from the menu.
    Kind regards,
       - Andreas.
    Edited by: Andreas Buchen on Feb 14, 2008 6:23 PM

  • Sql server service crashes a few times yesterday, however, there is no .MDMP file generated in the log folder

    hi there:
     I've experienced several sql server service crashes yesterday, service was brought up after a few secs as we have an autostart feature. 
    Today, I'd like to dig more into the errors and on the event viewer -> Application, I saw errors below
    "Faulting application sqlservr.exe, version 2009.100.4302.0, time stamp 0x52f5d194, faulting module ntdll.dll, version 6.0.6001.18538, time stamp 0x4cb73957, exception code 0xc0000374, fault offset 0x00000000000a7857, process id 0x15ec, application
    start time 0x01cfd7e72b2997d3.
    People suggested to check dump file inside C:\Program Files\Microsoft SQL Server\MSSQL10_50.SQL2008R2\MSSQL\Log
    However, the lastest .mdmp file was  on April 11, 2014. Do I need to turn on some settings on sql server in order to have this dump file? 
    Thanks
    Hui
    --Currently using Reporting Service 2000; Visual Studio .NET 2003; Visual Source Safe SSIS 2008 SSAS 2008, SVN --

    Results after sp_readerrorlog 1
    LogDate ProcessInfo
    Text
    2014-09-24 04:03:28.460 Server
    Microsoft SQL Server 2008 R2 (SP2) - 10.50.4302.0 (X64) 
    Feb  7 2014 17:23:24 
    Copyright (c) Microsoft Corporation
    Enterprise Edition (64-bit) on Windows NT 6.0 <X64> (Build 6001: Service Pack 1)
    2014-09-24 04:03:28.460 Server
    (c) Microsoft Corporation.
    2014-09-24 04:03:28.460 Server
    All rights reserved.
    2014-09-24 04:03:28.460 Server
    Server process ID is 5612.
    2014-09-24 04:03:28.460 Server
    System Manufacturer: 'IBM', System Model: 'IBM System x3650 -[7979AC1]-'.
    2014-09-24 04:03:28.460 Server
    Authentication mode is MIXED.
    2014-09-24 04:03:28.460 Server
    Logging SQL Server messages in file 'C:\Program Files\Microsoft SQL Server\MSSQL10_50.SQL2008R2\MSSQL\Log\ERRORLOG'.
    2014-09-24 04:03:28.460 Server
    This instance of SQL Server last reported using a process ID of 6780 at 9/24/2014 4:03:20 AM (local) 9/24/2014 11:03:20 AM (UTC). This is an informational message only; no user action is required.
    2014-09-24 04:03:28.460 Server
    Registry startup parameters: 
    -d C:\Program Files\Microsoft SQL Server\MSSQL10_50.SQL2008R2\MSSQL\DATA\master.mdf
    -e C:\Program Files\Microsoft SQL Server\MSSQL10_50.SQL2008R2\MSSQL\Log\ERRORLOG
    -l C:\Program Files\Microsoft SQL Server\MSSQL10_50.SQL2008R2\MSSQL\DATA\mastlog.ldf
    2014-09-24 04:03:28.480 Server
    SQL Server is starting at normal priority base (=7). This is an informational message only. No user action is required.
    2014-09-24 04:03:28.480 Server
    Detected 4 CPUs. This is an informational message; no user action is required.
    2014-09-24 04:03:28.480 Server
    Cannot use Large Page Extensions:  lock memory privilege was not granted.
    2014-09-24 04:03:28.580 Server
    Using dynamic lock allocation.  Initial allocation of 2500 Lock blocks and 5000 Lock Owner blocks per node.  This is an informational message only.  No user action is required.
    2014-09-24 04:03:28.700 Server
    Node configuration: node 0: CPU mask: 0x000000000000000f:0 Active CPU mask: 0x000000000000000f:0. This message provides a description of the NUMA configuration for this computer. This is an informational message only. No user action is required.
    2014-09-24 04:03:28.870 spid7s
    Starting up database 'master'.
    2014-09-24 04:03:28.940 spid7s
    1 transactions rolled forward in database 'master' (1). This is an informational message only. No user action is required.
    2014-09-24 04:03:28.940 spid7s
    0 transactions rolled back in database 'master' (1). This is an informational message only. No user action is required.
    2014-09-24 04:03:28.940 spid7s
    Recovery is writing a checkpoint in database 'master' (1). This is an informational message only. No user action is required.
    2014-09-24 04:03:29.040 spid7s
    CHECKDB for database 'master' finished without errors on 2014-09-20 12:00:01.697 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:29.050 spid7s
    Resource governor reconfiguration succeeded.
    2014-09-24 04:03:29.050 spid7s
    SQL Server Audit is starting the audits. This is an informational message. No user action is required.
    2014-09-24 04:03:29.050 spid7s
    SQL Server Audit has started the audits. This is an informational message. No user action is required.
    2014-09-24 04:03:29.050 spid7s
    FILESTREAM: effective level = 0, configured level = 0, file system access share name = 'SQL2008R2'.
    2014-09-24 04:03:29.070 spid7s
    SQL Trace ID 1 was started by login "sa".
    2014-09-24 04:03:29.070 spid7s
    Starting up database 'mssqlsystemresource'.
    2014-09-24 04:03:29.090 spid7s
    The resource database build version is 10.50.4302. This is an informational message only. No user action is required.
    2014-09-24 04:03:29.600 spid10s
    Starting up database 'model'.
    2014-09-24 04:03:29.600 spid7s
    Server name is 'BPWDB046\SQL2008R2'. This is an informational message only. No user action is required.
    2014-09-24 04:03:29.880 spid13s
    A new instance of the full-text filter daemon host process has been successfully started.
    2014-09-24 04:03:29.960 spid14s
    Starting up database 'msdb'.
    2014-09-24 04:03:29.960 spid17s
    Starting up database 'DW_Detail'.
    2014-09-24 04:03:29.960 spid16s
    Starting up database 'ReportServer_MSSQLSSRSTempDB'.
    2014-09-24 04:03:29.960 spid13s
    Starting up database 'Billing_Rep'.
    2014-09-24 04:03:29.960 spid15s
    Starting up database 'ReportServer_MSSQLSSRS'.
    2014-09-24 04:03:29.960 spid18s
    Starting up database 'BCBio_DW'.
    2014-09-24 04:03:29.960 spid19s
    Starting up database 'ODS'.
    2014-09-24 04:03:29.960 spid20s
    Starting up database 'ReportServer$SQL2008'.
    2014-09-24 04:03:29.970 spid21s
    Starting up database 'ReportServer$SQL2008TempDB'.
    2014-09-24 04:03:29.970 spid22s
    Starting up database 'Billing'.
    2014-09-24 04:03:29.970 spid23s
    Starting up database 'Billing_Snap'.
    2014-09-24 04:03:29.970 spid24s
    Starting up database 'DBA'.
    2014-09-24 04:03:30.000 Server
    A self-generated certificate was successfully loaded for encryption.
    2014-09-24 04:03:30.000 Server
    Server is listening on [ 'any' <ipv6> 52328].
    2014-09-24 04:03:30.000 Server
    Server is listening on [ 'any' <ipv4> 52328].
    2014-09-24 04:03:30.000 Server
    Server local connection provider is ready to accept connection on [ \\.\pipe\SQLLocal\SQL2008R2 ].
    2014-09-24 04:03:30.000 Server
    Server local connection provider is ready to accept connection on [ \\.\pipe\MSSQL$SQL2008R2\sql\query ].
    2014-09-24 04:03:30.010 Server
    Server is listening on [ ::1 <ipv6> 52329].
    2014-09-24 04:03:30.010 Server
    Server is listening on [ 127.0.0.1 <ipv4> 52329].
    2014-09-24 04:03:30.010 Server
    Dedicated admin connection support was established for listening locally on port 52329.
    2014-09-24 04:03:30.060 spid10s
    CHECKDB for database 'model' finished without errors on 2014-09-20 12:00:17.583 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:30.060 spid10s
    Clearing tempdb database.
    2014-09-24 04:03:30.140 Server
    The SQL Server Network Interface library successfully registered the Service Principal Name (SPN) [ MSSQLSvc/BPWDB046.bcbio.org:SQL2008R2 ] for the SQL Server service. 
    2014-09-24 04:03:30.140 Server
    The SQL Server Network Interface library successfully registered the Service Principal Name (SPN) [ MSSQLSvc/BPWDB046.bcbio.org:52328 ] for the SQL Server service. 
    2014-09-24 04:03:30.140 Server
    SQL Server is now ready for client connections. This is an informational message; no user action is required.
    2014-09-24 04:03:30.480 spid21s
    CHECKDB for database 'ReportServer$SQL2008TempDB' finished without errors on 2014-09-20 16:13:10.043 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:30.520 spid14s
    2933 transactions rolled forward in database 'msdb' (4). This is an informational message only. No user action is required.
    2014-09-24 04:03:30.650 spid7s
    0 transactions rolled back in database 'msdb' (4). This is an informational message only. No user action is required.
    2014-09-24 04:03:30.650 spid7s
    Recovery is writing a checkpoint in database 'msdb' (4). This is an informational message only. No user action is required.
    2014-09-24 04:03:30.770 spid14s
    CHECKDB for database 'msdb' finished without errors on 2014-09-20 12:00:18.473 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:30.950 spid10s
    Starting up database 'tempdb'.
    2014-09-24 04:03:30.990 spid52
    Attempting to load library 'xpstar.dll' into memory. This is an informational message only. No user action is required.
    2014-09-24 04:03:31.010 spid52
    Using 'xpstar.dll' version '2009.100.1600' to execute extended stored procedure 'xp_sqlagent_monitor'. This is an informational message only; no user action is required.
    2014-09-24 04:03:31.060 spid14s
    The Service Broker protocol transport is disabled or not configured.
    2014-09-24 04:03:31.060 spid14s
    The Database Mirroring protocol transport is disabled or not configured.
    2014-09-24 04:03:31.070 spid14s
    Service Broker manager has started.
    2014-09-24 04:03:31.330 spid16s
    CHECKDB for database 'ReportServer_MSSQLSSRSTempDB' finished without errors on 2014-09-20 12:00:39.873 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:31.500 spid20s
    2 transactions rolled forward in database 'ReportServer$SQL2008' (11). This is an informational message only. No user action is required.
    2014-09-24 04:03:31.960 spid7s
    0 transactions rolled back in database 'ReportServer$SQL2008' (11). This is an informational message only. No user action is required.
    2014-09-24 04:03:31.960 spid7s
    Recovery is writing a checkpoint in database 'ReportServer$SQL2008' (11). This is an informational message only. No user action is required.
    2014-09-24 04:03:32.150 spid20s
    CHECKDB for database 'ReportServer$SQL2008' finished without errors on 2014-09-20 16:13:07.583 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:33.170 spid24s
    CHECKDB for database 'DBA' finished without errors on 2014-09-20 16:22:41.603 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:33.600 spid13s
    Recovery of database 'Billing_Rep' (8) is 0% complete (approximately 1537 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:03:34.820 spid15s
    CHECKDB for database 'ReportServer_MSSQLSSRS' finished without errors on 2014-09-20 12:00:32.067 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:48.290 spid13s
    Recovery of database 'Billing_Rep' (8) is 1% complete (approximately 1488 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:03:49.520 spid17s
    1 transactions rolled forward in database 'DW_Detail' (7). This is an informational message only. No user action is required.
    2014-09-24 04:03:50.070 spid7s
    0 transactions rolled back in database 'DW_Detail' (7). This is an informational message only. No user action is required.
    2014-09-24 04:03:50.070 spid7s
    Recovery is writing a checkpoint in database 'DW_Detail' (7). This is an informational message only. No user action is required.
    2014-09-24 04:03:50.070 spid7s
    Recovery completed for database DW_Detail (database ID 7) in 5 second(s) (analysis 251 ms, redo 4314 ms, undo 44 ms.) This is an informational message only. No user action is required.
    2014-09-24 04:03:50.520 spid22s
    CHECKDB for database 'Billing' finished without errors on 2013-04-25 18:47:20.743 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:03:51.620 spid17s
    CHECKDB for database 'DW_Detail' finished without errors on 2014-09-20 12:00:46.917 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:04:03.740 spid13s
    Recovery of database 'Billing_Rep' (8) is 2% complete (approximately 1493 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:13.040 spid13s
    Recovery of database 'Billing_Rep' (8) is 0% complete (approximately 8052 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:23.220 spid18s
    Database BCBio_DW has more than 1000 virtual log files which is excessive. Too many virtual log files can cause long startup and backup times. Consider shrinking the log and using a different growth increment to reduce the number of virtual log files.
    2014-09-24 04:04:23.420 spid18s
    Recovery of database 'BCBio_DW' (9) is 0% complete (approximately 4664 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:25.640 spid18s
    Recovery of database 'BCBio_DW' (9) is 0% complete (approximately 4954 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:25.640 spid18s
    Recovery of database 'BCBio_DW' (9) is 3% complete (approximately 74 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:25.730 spid18s
    Recovery of database 'BCBio_DW' (9) is 2% complete (approximately 98 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:26.000 spid18s
    Recovery of database 'BCBio_DW' (9) is 12% complete (approximately 18 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:28.000 spid19s
    Database ODS has more than 1000 virtual log files which is excessive. Too many virtual log files can cause long startup and backup times. Consider shrinking the log and using a different growth increment to reduce the number of virtual log files.
    2014-09-24 04:04:28.350 spid19s
    CHECKDB for database 'ODS' finished without errors on 2014-09-20 16:07:22.603 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:04:29.370 spid18s
    Recovery of database 'BCBio_DW' (9) is 19% complete (approximately 24 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:31.380 spid18s
    Recovery of database 'BCBio_DW' (9) is 25% complete (approximately 22 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:32.490 spid18s
    Recovery of database 'BCBio_DW' (9) is 32% complete (approximately 18 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:33.150 spid13s
    Recovery of database 'Billing_Rep' (8) is 0% complete (approximately 9805 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:33.860 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 04:04:33.860 Logon
    Login failed for user 'ETLAdmin'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.101.33]
    2014-09-24 04:04:34.040 spid18s
    Recovery of database 'BCBio_DW' (9) is 39% complete (approximately 15 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:35.600 spid18s
    Recovery of database 'BCBio_DW' (9) is 46% complete (approximately 13 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:36.740 spid18s
    Recovery of database 'BCBio_DW' (9) is 53% complete (approximately 10 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:37.220 spid13s
    Recovery of database 'Billing_Rep' (8) is 0% complete (approximately 10106 seconds remain). Phase 1 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:37.230 spid13s
    Recovery of database 'Billing_Rep' (8) is 2% complete (approximately 2587 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:37.360 spid13s
    Recovery of database 'Billing_Rep' (8) is 1% complete (approximately 3654 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:37.560 spid13s
    Recovery of database 'Billing_Rep' (8) is 3% complete (approximately 1564 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:38.070 spid13s
    Recovery of database 'Billing_Rep' (8) is 4% complete (approximately 1328 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:38.830 spid18s
    Recovery of database 'BCBio_DW' (9) is 61% complete (approximately 8 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:40.080 spid13s
    Recovery of database 'Billing_Rep' (8) is 5% complete (approximately 1268 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:40.680 spid18s
    Recovery of database 'BCBio_DW' (9) is 69% complete (approximately 6 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:42.960 spid18s
    Recovery of database 'BCBio_DW' (9) is 77% complete (approximately 4 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:44.870 spid18s
    Recovery of database 'BCBio_DW' (9) is 85% complete (approximately 3 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:44.870 spid18s
    3 transactions rolled forward in database 'BCBio_DW' (9). This is an informational message only. No user action is required.
    2014-09-24 04:04:45.120 spid18s
    Recovery of database 'BCBio_DW' (9) is 85% complete (approximately 3 seconds remain). Phase 3 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:04:46.870 spid7s
    0 transactions rolled back in database 'BCBio_DW' (9). This is an informational message only. No user action is required.
    2014-09-24 04:04:46.870 spid7s
    Recovery is writing a checkpoint in database 'BCBio_DW' (9). This is an informational message only. No user action is required.
    2014-09-24 04:04:46.870 spid7s
    Recovery completed for database BCBio_DW (database ID 9) in 23 second(s) (analysis 2417 ms, redo 19227 ms, undo 1745 ms.) This is an informational message only. No user action is required.
    2014-09-24 04:04:47.610 spid18s
    CHECKDB for database 'BCBio_DW' finished without errors on 2014-09-20 14:02:10.680 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:04:47.820 spid23s
    Database Billing_Snap has more than 1000 virtual log files which is excessive. Too many virtual log files can cause long startup and backup times. Consider shrinking the log and using a different growth increment to reduce the number of virtual log files.
    2014-09-24 04:04:48.570 spid23s
    CHECKDB for database 'Billing_Snap' finished without errors on 2014-09-20 16:13:11.037 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:04:52.970 spid13s
    Recovery of database 'Billing_Rep' (8) is 6% complete (approximately 1247 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:05:05.540 spid13s
    Recovery of database 'Billing_Rep' (8) is 7% complete (approximately 1224 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:05:17.980 spid13s
    Recovery of database 'Billing_Rep' (8) is 8% complete (approximately 1202 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:05:22.450 spid13s
    Recovery of database 'Billing_Rep' (8) is 8% complete (approximately 1196 seconds remain). Phase 2 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:05:22.450 spid13s
    1 transactions rolled forward in database 'Billing_Rep' (8). This is an informational message only. No user action is required.
    2014-09-24 04:05:22.560 spid13s
    Recovery of database 'Billing_Rep' (8) is 8% complete (approximately 1196 seconds remain). Phase 3 of 3. This is an informational message only. No user action is required.
    2014-09-24 04:05:22.710 spid13s
    CHECKDB for database 'Billing_Rep' finished without errors on 2014-09-20 13:25:01.597 (local time). This is an informational message only; no user action is required.
    2014-09-24 04:05:22.720 spid13s
    0 transactions rolled back in database 'Billing_Rep' (8). This is an informational message only. No user action is required.
    2014-09-24 04:05:22.720 spid13s
    Recovery is writing a checkpoint in database 'Billing_Rep' (8). This is an informational message only. No user action is required.
    2014-09-24 04:05:22.730 spid13s
    Recovery completed for database Billing_Rep (database ID 8) in 109 second(s) (analysis 63983 ms, redo 45211 ms, undo 154 ms.) This is an informational message only. No user action is required.
    2014-09-24 04:05:22.750 spid7s
    Recovery is complete. This is an informational message only. No user action is required.
    2014-09-24 07:00:06.930 spid53
    Common language runtime (CLR) functionality initialized using CLR version v2.0.50727 from C:\Windows\Microsoft.NET\Framework64\v2.0.50727\.
    2014-09-24 07:00:07.720 spid53
    AppDomain 2 (mssqlsystemresource.sys[runtime].1) created.
    2014-09-24 07:13:38.590 spid51
    SQL Server blocked access to STATEMENT 'OpenRowset/OpenDatasource' of component 'Ad Hoc Distributed Queries' because this component is turned off as part of the security configuration for this server. A system administrator can enable the use of 'Ad
    Hoc Distributed Queries' by using sp_configure. For more information about enabling 'Ad Hoc Distributed Queries', see "Surface Area Configuration" in SQL Server Books Online.
    2014-09-24 07:13:53.870 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 07:13:53.870 Logon
    Login failed for user 'BCBIO\sqlservice'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.100.6]
    2014-09-24 07:13:56.560 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 07:13:56.560 Logon
    Login failed for user 'BCBIO\sqlservice'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.100.6]
    2014-09-24 07:13:57.460 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 07:13:57.460 Logon
    Login failed for user 'BCBIO\sqlservice'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.100.6]
    2014-09-24 07:13:57.560 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 07:13:57.560 Logon
    Login failed for user 'BCBIO\sqlservice'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.100.6]
    2014-09-24 07:13:57.950 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 07:13:57.950 Logon
    Login failed for user 'BCBIO\sqlservice'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.100.6]
    2014-09-24 07:14:04.410 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 07:14:04.410 Logon
    Login failed for user 'BCBIO\sqlservice'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.100.6]
    2014-09-24 07:14:04.960 Logon
    Error: 18456, Severity: 14, State: 38.
    2014-09-24 07:14:04.960 Logon
    Login failed for user 'BCBIO\sqlservice'. Reason: Failed to open the explicitly specified database. [CLIENT: 172.20.100.6]
    2014-09-24 07:34:13.330 spid54
    Error: 17053, Severity: 16, State: 1.
    2014-09-24 07:34:13.330 spid54
    V:\log\templog.ldf: Operating system error 112(There is not enough space on the disk.) encountered.
    2014-09-24 07:34:14.430 spid54
    Error: 9002, Severity: 17, State: 4.
    2014-09-24 07:34:14.430 spid54
    The transaction log for database 'tempdb' is full. To find out why space in the log cannot be reused, see the log_reuse_wait_desc column in sys.databases
    --Currently using Reporting Service 2000; Visual Studio .NET 2003; Visual Source Safe SSIS 2008 SSAS 2008, SVN --

  • Request - AI files copy & Curved in different folder

    Dear Participants...
    Is it possible with javascript code to do the following?
    Have AI files from different paths. I need a copy all files with text curves. Should work for batch.
    In the given path one folder should cretae called 'Curved" and all the files need copy to that folder and add file name at end '_Curved'
    For example: file name is 'Filename.ai'. It should come 'Filename_Curved.ai'
    Using Adobe CS4
    if you help me that would be really grateful...
    Thanks in advance..

    I prefer to script than work… Anyhow's I gave this a quick test here and it works fine for me…
    #target illustrator
    var df = Folder('~/Desktop');
    var aiFolder = Folder.selectDialog('Please choose your AI Folder…', df);
    if (aiFolder != null) {
         fileList = aiFolder.getFiles(/\.ai$/i);
         if (fileList.length > 0) {
              curveFolder = Folder(aiFolder.fsName+'/Curved')
              if (!curveFolder.exists) curveFolder.create();
              aiProcess();
         } else {
              alert('This Folder contained NO Illustrator AI files?');
    function aiProcess() {
         var uIL = app.userInteractionLevel;
         app.userInteractionLevel = UserInteractionLevel.DONTDISPLAYALERTS;
         for (var i = 0; i < fileList.length; i++) {
              app.open(fileList[i]);
              var docRef = app.activeDocument;
              recurseLayers(docRef.layers);
              var docName = docRef.name.replace(/\.ai$/i,'');
              aiFilePath = File(curveFolder.fsName+'/'+docName+'_Curved.ai');
              var aiOpts = saveAsAiFile();
              docRef.saveAs(aiFilePath, aiOpts);
              docRef.close(SaveOptions.DONOTSAVECHANGES);
         app.userInteractionLevel = uIL;
    function saveAsAiFile() {
         var aiOpts = new IllustratorSaveOptions();
         with (aiOpts) {
              compatibility = Compatibility.ILLUSTRATOR12;
              compressed = true;
              embedICCProfile = true;
              embedLinkedFiles = true;
              flattenOutput = OutputFlattening.PRESERVEAPPEARANCE;
              fontSubsetThreshold = 0;
              overprint = PDFOverprint.PRESERVEPDFOVERPRINT;
              pdfCompatible = true;
         return aiOpts;
    function recurseLayers(objArray) {
         for (var i = 0; i < objArray.length; i++) {
              var l = objArray[i].locked;
              if (l) objArray[i].locked = false;
              var v = objArray[i].visible;
              if (!v) objArray[i].visible = true;
              outlineText(objArray[i].textFrames);
              if (objArray[i].layers.length > 0) {
                   recurseLayers(objArray[i].layers)
              if (objArray[i].groupItems.length > 0) {
                   recurseGroups(objArray[i].groupItems)
              objArray[i].locked = l;
              objArray[i].visible = v;
    function recurseGroups(objArray) {
         for (var i = 0; i < objArray.length; i++) {
              var l = objArray[i].locked;
              if (l) objArray[i].locked = false;
              var h = objArray[i].hidden;
              if (h) objArray[i].hidden = false;
              outlineText(objArray[i].textFrames);
              if (objArray[i].groupItems.length > 0) {
                   recurseGroups(objArray[i].groupItems)
              objArray[i].locked = l;
              objArray[i].hidden = h;
    function outlineText(objArray) {
         for (var i = objArray.length-1; i >= 0; i--) {
              var l = objArray[i].locked;
              if (l) objArray[i].locked = false;
              var h = objArray[i].hidden;
              if (h) objArray[i].hidden = false;
              var g = objArray[i].createOutline();
              g.locked = l;
              g.hidden = h;    
    There is a line '61' in my ESTK where you may want to change the version number it would be '15' for CS5…

  • How to verify dump file

    Is there any way to verify dump file generated by export utility ?
    Can I use the dbverify ?
    Thanks

    dmp file can only be read by imp utility, not other one. dbverify is for different purpose, not at least for the one you ask.
    What you can could do is, just import the dmp file to know whether it is real dmp file or not.
    SJH
    OCP DBA

  • Redirect core dump file

    This seems like a fairly simple problem, but I cannot seem to find it in any searches...
    How can I set the core (or heap) dump directory to "/usr/scratch" instead of the current directory. I running 1.4.2 for now.
    Thanks for any help!

    Hi, Im not familiar with heap dumps so i apologize if this isnt helpful...
    http://blogs.sun.com/roller/page/alanb?entry=heap_dumps_are_back_with
    If you don't want big dump files in the application working directory then the HeapDumpPath option can be used to specify an alternative location - for example -XX:HeapDumpPath=/disk2/dumps will cause the heap dump to be generated in the /disk2/dumps directory.
    https://hat.dev.java.net/doc/README.html
    For example, to run the program Main and generate a binary heap profile in the file dump.hprof, use the following command:
    java -Xrunhprof:file=dump.hprof,format=b Main
    also:
    http://www.hp.com/products1/unix/java/infolibrary/prog_guide/hotspot.html
    -XX:+HeapDumpOnOutOfMemory
    The HeapDumpOnOutOfMemory command line option causes the JVM to dump a snapshot of the Java heap when an Out Of Memory error condition has been reached. The heap dump format generated by HeapDumpOnOutOfMemory is in hprof binary format, and is written to filename java_pid<pid>_<.hprof in the current working directory. The option -XX:HeapDumpPath=<file>_< can be used to specify the dump filename or the directory where the dump file is created. Running with application with -XX:+HeapDumpOnOutOfMemoryError does not impact performance. Please note the following known issue: The HeapDumpOnOutOfMemory option does not work with the low-pause collector (-XX:+UseConcMarkSweepGC). This option is available starting with the 1.4.2.11 and 1.5.0.04 releases.

  • How to access beans from different folder

    hi guys,
    currently i have different folders in the webapps folder found in Tomcat 4.1. I was asked to centralised the codes(that is used in all JSP file) so that it will be easier for future maintainence. After realising that I cant use the jsp:include tag, I have plan to use JavaBean so that whenever the jsp page is called it will access the JavaBean to retrieve the arguments. I cannot retrieve the arguments from the URL directly because the JSP files are embedded in a portlet.
    So is there any way I can read the JavaBean when my jsp file is in the different folder???
    *Each folder in the webapps directory is a seperate context/web application.
    Any help is greatly n deeply deeply appreciated.

    As u have said i have to write myself (store data
    accessed across applications). Wat do you mean by
    that? Isit I have to store the values in a table
    first?Yes, that's the general idea. You will have to setup a db and
    some tables for shared data across applications (say one user signs
    on to one application and his credentials should be carried over when
    he accesses the other application via a link in the first). The
    implementation is very dependent on what it is that needs to be shared
    and other requirements around it (like, if its required to have a timeout
    value and so on). I have never done such stuff myself, but know of
    projects which do so quite succesfully.
    cheers,
    ram.

  • Screenshot Error in Mac OS X 10.8, Short key working but no file generated

    Today I'm going to use the mac screenshot (screen capture from COMMAND+CONTROL+SHIFT+4) to capture the window while the problem comes. The short key is working and the icon of pointer has changed accordingly, but once I clicked there is no file generated in the destination folder. I would like to know what's wrong with my screenshot? Is there anyone who encountered the similar errors before?
    I'm using Mac OS X 10.8.5 and I believe the screenshot is still avalaible after the update from the 10.7.x. I guess the problem is more about the screenshot applet rather than the system. Besides, I'm eager to know that is there any method to solve this problem.
    Thank you.

    I am running a Retina Macbook Pro with Mavericks 10.9.3 installed.  I was having this problem today, but I did a few things that got rid of the issue...
    1)  Go to System Preferences > Keyboard > Shortcuts > Screen Shots, and click each of these 4 boxes...
    Then close the window.
    2) Open the application "Terminal."
    Type in killall Finder > Push enter
    Type in killall Dock > Push enter
    Close the program
    3) Turn off the computer.  Now turn it on, but hold down "CMD + OPTION + P + R" on the keyboard as it boots up.  This will restart your computer.  Let the keys go after the restart, and let it boot all the way.  Pressing these keys clears your PRAM and NVRAM settings--it doesn't harm your computer at all, but it is supposed to help give your battery a boost.  Apparently, it also resolved my screen shot issues.
    I blame my screen shot issues on a program I tried installing today--it must have messed with the OS somehow.  Hope all of this helps!

  • To Create Time stamp in dump file during export

    Hi All ,
    While exporting table thDIRECTORY=USER_DATA_PUMP_DIR
    DUMPFILE=expdp_tables%U.dmp
    LOGFILE=expdp_tables.log
    CONTENT=ALL
    PARALLEL=4
    ESTIMATE=STATISTICS
    SCHEMAS=<Schema name >
    When i ran the shell file i am getting dump file generated as
    expdp_tables01.dmp
    expdp_tables02.dmp
    Could any one help how can i get the time stamp attached to the dump file .Its urgent kindly help me in this
    Thanks

    Munich,
    Thanks for ur suggestions and the same need to be imported how can i do it in the import operation while my import par file has
    DIRECTORY=DATA_PUMP_DIR
    DUMPFILE=expdp_tables%U.dmp <<how can i get the sydate.dmp>>
    LOGFILE=impdp_EIM_tables.log
    CONTENT=DATA_ONLY
    PARALLEL=4
    TABLE_EXISTS_ACTION=TRUNCATE
    EXCLUDE=GRANT
    EXCLUDE=STATISTICS
    EXCLUDE=TRIGGER
    What are the option that i can give to get the sysdate.dmp in the par file that i have got from
    mv exp.dmp exp.dmp.`date "+%d.%m.%Y"`
    or can i give
    DUMPFILE=expdp_tables%U'date'.dmp like that
    Thanks

  • Very Large Heap Dump

    Hi,
    I've to analyze a huge heap dump file (ca. 1GB).
    I've tried some tools HAT, YourKit, Profiler4J... etc.)
    Always an OutOfMemoryError occures.
    My machine has 2GB physocal Memory and 3GB SwapSpace on a Dual Core Intel Processor with 2,6GHz
    Is there a way to load the file on my machine or is there a tool which is able to load dump files partially?
    ThanX ToX

    1GB isn't very large :-) When you say you tried HAT, did you mean jhat? Just checking as jhat requires less memory than the original HAT. Also, did you include an option such as -J-mx1500m to increase the maximum heap for the tools that you tried? Another one to try is the HeapWalker tool in VisualVM. That does a better job than HAT/jhat with bigger heaps.

  • How to generate & publish  webhelp html files in different folder than default folder

    Hi All,
    I am new to robo help. I have a requirement to generate & publish webhelp html files in a different folder than default folder.
    Could anyone please tell me the steps to change the folder
    Thanks
    Rashmi

    You change the generate folder and filename in the first field on the first page of the wizard. It must be a folder on your hard disk.
    You change the publish folder in the last page of the wizard. Anywhere you like.
    See www.grainge.org for RoboHelp and Authoring tips
    @petergrainge

  • Issue in getting exception pointer in VC++ CLR project for generating dump file.

    Hi,
    We are working on VC++ CLR (Windows Forms) application.
    We are writing a code to generate Mini dump file (using MINIDUMPWRITEDUMP in dbghelp.dll) whenever application crashes.
    The code available on internet for writing mini dump file in c# generates the mini dump but without Exception code and information, because _EXCEPTION_INFORMATION was set using Marshal::GetExceptionPointers() which returns 0. We called MINIDUMPWRITEDUMP
    inside AppDomain::UnhandledExceptionEventHandler for this.
    To get the valid exception pointer for mini dump we made following attempts with and without AppDomain::UnhandledExceptionEventHandler  –
    Used mixed (managed and unmanaged ) dll to add SetUnhandledExceptionFilter.  Inside top level exception filter function added code to write mini dump.  Inside main the handler is set on 1<sup>st</sup> line. But handler is not invoked
    on any exception. This works fine with console C# project.
    // inside dll
    #pragma unmanaged
    LONG WINAPI ABCExceptionFilter(__in struct _EXCEPTION_POINTERS *pExceptionInfo)
    { // code to write dump
    #pragma managed
    Namespacee SEHHandler {
    public ref class SEHExpHandler
    public:
    static void AddHandler()
    SetUnhandledExceptionFilter(ABCExceptionFilter);
    // inside app main()
    [STAThreadAttribute]
    int main(array<System::String ^> ^args)
    SEHHandler::SEHExpHandler::AddHandler();
    //some code . . .
    Application::Run(gcnew Form1());
    //some code . . .
    Used __try __except inside application main. The exception filter was written in other C++ dll and called inside __except ( expression ) using platform Invoke.  But this throws the caught exception while trying to call handler.
    // inside app class
    [DllImport("SEHhandler.dll")]
    extern int __stdcall se_handler(unsigned int code, IntPtr ep);
    [STAThreadAttribute]
    int main(array<System::String ^> ^args)
    __try
    { //some code. . .
    __except(se_handler(GetExceptionCode(),
    (IntPtr)GetExceptionInformation()))
    { //some code. . .
    // inside dll
    __declspec(dllexport)
    extern int WINAPI se_handler(unsigned int code,
    struct _EXCEPTION_POINTERS* ep)
    { // code to write minidump
    return EXCEPTION_EXECUTE_HANDLER;
    Can anyone help us to get either of the code work. The aim is to write dump file with correct stack trace where exception was thrown.
    Thanks in advance.
    -- Sumit

    Hi,
    Few days back we tried following sample program and we were able to get exception code and pointer for almost all types of exception catch inside __except statement. 
    1. Create sample VC++ > CLR > Windows form application.
    2. Change Common language Runtime support to (/clr).
    3. Set UnhandledExceptionMode to ThrowException in main.
    4. Inside application cpp file add __try __except around Application::Run(gcnew Form1()).
    5. Add new header file to write unmanaged function to be called inside __except statement to handle exception.
    6. Created new Form1 and add few buttons to test different types of exceptions.
    7. Exceptions will be catch in __except statement and exception pointers will be available to write mini dump.
    Inside application main function use __try, __except block.
    #include "stdafx.h"
    #include "Form1.h"
    [STAThreadAttribute]
    int main(array<System::String ^> ^args)
    Application::SetUnhandledExceptionMode(UnhandledExceptionMode::ThrowException);
    Application::EnableVisualStyles();
    Application::SetCompatibleTextRenderingDefault(false);
    __try{
    // Create the main window and run it
    Application::Run(gcnew Form1());
    }__except(exception_filter(GetExceptionCode(), GetExceptionInformation())){
    // write your code
    return 0;
    Add header file and write exception filter
    #pragma once
    #include <windows.h>
    #include <stdio.h>
    #include <string>
    using namespace std;
    #pragma comment(lib, "user32.lib")
    int exception_filter(unsigned int code, struct _EXCEPTION_POINTERS* ep)
    TCHAR buff[100];
    wsprintf(buff, L"Inside exception_filter. Excp Pointer %p", (void*)ep);
    ::MessageBox(NULL,buff,L"Error",MB_OK);
    // add code to write mini dump using MINIDUMPWRITEDUMP
    return EXCEPTION_EXECUTE_HANDLER;
    Thanks.

Maybe you are looking for

  • There are many rt.jar

    Question 1: ========== - I started with JDK1.2.2 but lately installed JBUILDER development tools that has JDK1.3 I assume I could delete the 1.2.2 files - I re-adjusted the PATH env variable to contain c:\jbuilder\jdk1.3\bin instead of the previous c

  • Power calibration area error (0x73, 0x03.) ??? HELP!

    At the beginning of iDVD writing the lead-in on a burn, I am encountering "errors were found during the burning process". The explanation is- "The recording device reported the media error: Power calibration area error. (0x73, 0x03.)" Then I get a se

  • Problem connecting to new laptop

    I recently purchased a new laptop with windows 8 on, now i find i cant sync my ipod up with the new laptop everytime i do it comes up withcannot sync unknown error (-50).   ive tried putting it back on my old laptop because other sites have told me t

  • How to display Caller Name and Company

    How can i change the display of the incomming caller details to include the company name, without using third party utilities ? Thanks Pete

  • In my hp windows 7 laptop the recovery (Back to factory settings) is disabled.

    in my hp windows 7 laptop, the recovery (Back to factory original settings) is disabled in my HP recovery manager. please help me to enable the recovery options to restore my laptop to original factery settings