Editing very large panorama

Hi all,
I work actually on a huge project, a gigapixel panorama. The picture will go over the photoshop limitation around 300'000px and will be around 394110px large.
I will save the picture in psb format, and try to do the maximum corrections in the source pictures, but i may need to do editing after the stitching process.
Did you know if it is possible to edit this picture by cutting it in more usefull size pictures?
I will not do contrast or light corrections than can affect the final image if i merge the tiles picture, but do small corrections like blur people face, car immatriculation etc...
Thanks by advance
Regards

Sure. Cutting up the picture into smaller pieces will make it easier to work with. Bert Monroy (do a google search) works in this way for some of his really large scale paintings. Of course, running in 64bit and having tons of RAM in your Mac will help your cause in a substantial way.

Similar Messages

  • Editing very large images

    I have several very large images that require minor editing. The tasks that I require are only basic rotation, cropping, and maybe minor color adjustments.
    The problem is that the images range from 20780 x 15000 px down to 11150 x 2600 px.
    With images this size, when I try to perform any operation on the entire image, like a 3 deg rotation, fireworks simply times out. Maybe it is because it can't generate the preview image at that size, maybe it simply can't calculate the image at that size. If it is just the preview image, is there a way to lower the quality (say, jpeg quality of 5%) of the preview image only but when I export the image, I can keep it at 100%?
    Thank you,
    -Adam
    (Yes, I did search, but the forums search seemed mildly useless and engine results consistently returned nothing)

    Fireworks is not designed to work with images of this size. It's a screen graphics application. Photoshop would be a better option here, or other software designed for working with high resolution files.
    Jim Babbage

  • Editing very large MP3 Files

    Hello,
    Sorry for my bad English! I am French!
    I recently bought a Sony digital recorder (ICD-PX312) that records MP3 files.
    I recorded my noisy neighbour, in continuous mode for three days. (I was not at home)
    PX312 recorder has produced several files which maximum size is  511,845 kB corresponding to a lenght of 24H 15min.
    Every file, checked with Mediainfo, have the same properties. (48kbps 44.1KHz)
    I can read them with VLC media player without any problem, but I need to edit these files to keep only the interesting parts.
    If I open (drag and drop or open file) these files with Audition 1.0 (I came from Cool-Edit), I found a very strange behavior.
    The 24H 15min files are opened with a lenght of 1H and a half or so!.
    I gave a try opening with Audition 3.0 and the result is strange too, and more "funny":
    The lenght of one 24H15min file is 1H and a half
    The lenght of another 24H 15min file is 21H or so.
    In the Audition TMP directory, Audition has created several 4 GB temporay files that correspond to the limit of the WAV files.
    I made a test with a 128kpbs, 44.1khz. This 511,845 kB file is  9H 5min long
    Audition read it as 4H40min
    The TMP file is 2,897,600 kB, far below the 4 GB limit
    It seems Audition 1 and 3 (I read CS5 has the same problem too) does not share the WAV conversion very well.
    Is it due to Audition itself or to the Fraunhofer MP3 codec?
    Is there any workaround from Adobe?
    As I am an AVS4YOU client, I tried AVS Audio Editor. It opens every file with the good lenght but does not have the needed editing functions. That demonstrates that it is possible to share very large files.
    Many thanks in advance for any help or idea, because Audition is the editor I need!
    Best Regards

    SteveG wrote :
    t's not their 'bug' at all. MP3 files work fine, and Adobe didn't invent the wav file format, Microsoft did. It's caused by the 32-bit unsigned integer to record the file size headerincorporated into wav files, which limits their size to one that can be described within that limit - 4Gb.
    May be I was not clear enough in my explanation
    I Agree partly with you.
    I Agree with you when you wrote that the 4 GB limit is inherant to the Microsoft WAV format.
    When reading 24H MP3 Files, Audition 3.0 creates, in TMP folder, two chunk of 4 GB and another smaller chunk, to try to overcome the limit. This cutting process is exactly what I am expecting from such a software.
    The problem - and the bug - is that the duration  that is "extracted" in audition is smaller than that of the original MP3 (e.g. 21H 30min instead of 24H 15min). Some part of the original MP3 files has been skipped in the cuttng/conversion process.
    I dont think Microsoft is involved in this "cutting" process that belongs to Adobe an/or Fraunhofer. This is why I am surprised

  • Help required to make very large panoramic print.

    I have a large panoramic picture which was stitched to gether from about 6 verticaly taken shots.  The image size is currently about 7500 x 2500 pixels.  This is almost exactly a ration of 3 wide to 1 high.
    I found a local shop today that says it can  can print it 63cm wide & any length I wish.  This suggests that the maximun size they are able to print would be 63 cm high by about 190 cm long.
    What sort of resampling would you suggest & what number of pixels would you think I would need to do this?

    The rule of thumb is that a resolution of 240-300 px/inch is desirable for a decent print. With a large print, the resolution at the low end of the range should be ok, since the observer's eye is at a distance for viewing same.
    One would like to think that the local shop would have guidelines based on experience with their equipment.
    Have you checked with other vendors, e.g.Shutterfly, Mpix ?

  • Today, I randomly happened to have less than 1GB of hard drive space left. I found very large "frame" files, what are they?

    I found very large "frame" files, what are they & can I delete them? (See screenshot). I'm a (17 today)-year-old film-maker and can't edit in FCP X anymore because I "don't have enough space". Every time I try to delete one, another identical file creates itself...
    If that can help: I just upgraded to FCP 10.0.4 and every time I launch it it asks to convert my current projects (I know it would do it at least once) and I accept, but everytime I have to get it done AGAIN. My computer is slower than ever and I have a deadline this friday
    I also just upgraded to Mac OS X 10.7.4, and the problem hasn't been here for long, so it may be linked...
    Please help me!
    Alex

    The first thing you should do is to back up your personal data. It is possible that your hard drive is failing. If you are using Time Machine, that part is already done.
    Then, I think it would be easiest to reformat the drive and restore. If you ARE using Time Machine, you can start up from your Leopard installation disc. At the first Installer screen, go up to the menu bar, and from the Utilities menu, first select to run Disk Utility. Completely erase the internal drive using the Erase tab; make sure you have the internal DRIVE (not the volume) selected in the sidebar, and make sure you are NOT erasing your Time Machine drive by mistake. After erasing, quit Disk Utility, and select the command to restore from backup from the same Utilities menu. Using that Time Machine volume restore utility, you can restore it to a time and date immediately before you went on vacation, when things were working.
    If you are not using Time Machine, you can erase and reinstall the OS (after you have backed up your personal data). After restarting from the new installation and installing all the updates using Software Update, you can restore your personal data from the backup you just made.

  • Capture Image Of A Very Large JPanel

    Below is some code used to save an image of a JPanel to a file...
        int w = panel.getSize().width;
        int h = panel.getSize().height;
        BufferedImage image = new BufferedImage(w, h, BufferedImage.TYPE_INT_RGB);
        Graphics graphics = image.getGraphics();
        // Make the component believe its visible and do its layout.
        panel.addNotify();
        panel.setVisible(true);
        panel.validate();
        // Draw the graphics.
        panel.print(graphics);
        // Write the image to a file.
        ImageFile imageFile = new ImageFile("test.png");
        imageFile.save(image);
        // Dispose of the graphics.
        graphics.dispose();This works fine but my problem is that I am trying to save what may be a very large JPanel, perhaps as large as 10000x10000 pixels. It doesn't take long for the java heap to be used up and an exception to be thrown.
    I know I can increase the heap size of the JVM but since I can't ever be sure how large the panel may be that's a far from ideal solution.
    So the question is how do I save an image of a very large JPanel to a file?

    1) Does the OoM happens while instantiating the buffered image, (which probably tries to allocate a big continuous native array of pixels)?
    Or the Graphics object (same reason, though the Graphics is probably just an empty shell over the big pixel array)?
    2) In which format do you need to save the image? Do you only need to be able to read it again in your own program?
    If yes to both questions, then a pulled-by-the-hair solution coud be to instantiate your own Graphics subclass (no Buffered Image), whose operations would save their arguments directly to the image file, instead of into a big in-memory model of the panel image.
    If the output format is a standard one though (GIF, JPG,...), then maybe your custom Graphics's operations could contain the logic to encode/compress as much as possible of the arguments into an in-memory bytearray of the target format?
    I'm not very confident though; I d'ont know the GIF or JPEG encoding, but I suspect (especially for JPEG) that you need to know the "whole" image to encode it properly.
    But if the target format supports encoders that work on the fly out of streams of bytes (e.g. BMP ) then you can use whatever compress/uncompress technique you see fit (e.g. RLE ): you know the nature of the panels, you may be aware of some optimizations you may perform wrt pixels storage. prior to encoding (e.g., bug empty areas, predictable chessboard pattern, black-and-white palette,...).
    Edited by: jduprez on Sep 19, 2009 7:33 PM

  • How can NI FBUS Monitor display very large recorded files

    NI FBUS Monitor version 3.0.1 outputs an error message "Out of memory", if I try to load a large recorded file of size 272 MB. Is there any combination of operating system (possible Vista32 or Vista64) and/or physical memory size, where NI FBUS Monitor can display such large recordings ? Are there any patches or workarounds or tools to display very large recorded files?

    Hi,
    NI-FBUS Monitor does not set the limitation on the maximum record file size. The physical memory size in the system is one of the most important factors that affect the loading of large record file.  Monitor will try loading the entire file into the memory during file open operation.
    272MB is a really large file size. To open the file, your system must have sufficient physical memory available.  Otherwise "Out of memory" error will occur.
    I would recommend you do not use Monitor to open a file larger than 100MB. Loading of a too large file will consume the system memory quickly and decrease the performance.
    Message Edited by Vince Shen on 11-30-2009 09:38 PM
    Feilian (Vince) Shen

  • When using reader, the font gets very large on last page and I cant change it

    when using reader, the font gets very large on last page and I cant change it

    You cannot edit PDF files with Reader. You can change the View options to make what you see smaller. View -> Zoom

  • Very large bdump file sizes, how to solve?

    Hi gurus,
    I currently always find my disk space is not enough, after checking, it is the oraclexe/admin/bdump, there's currently 3.2G for it, my database is very small, only holding datas of 10mb.
    It didn't happen before, only currently.
    I don't know why it happened, I have deleted some old files in that folder, but today I found it is still very large compare to my database.
    I am running an apex application with xe, the applcaitions works well, we didn't see anything wrong, but only the bdump file very big.
    any tip to solve this? thanks
    here comes my alert_xe.log file content:
    hu Jun 03 16:15:43 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5600.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:15:48 2010
    Restarting dead background process MMON
    MMON started with pid=11, OS id=5452
    Thu Jun 03 16:15:52 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:16:16 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:20:54 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:21:50 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:25:56 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:26:18 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:30:58 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:31:19 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:36:00 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5452.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:36:46 2010
    Restarting dead background process MMON
    MMON started with pid=11, OS id=1312
    Thu Jun 03 16:36:49 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:37:13 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:41:51 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:42:13 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:46:54 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:47:17 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:51:57 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:52:35 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:56:58 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1312.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:57:10 2010
    Restarting dead background process MMON
    MMON started with pid=11, OS id=3428
    Thu Jun 03 16:57:13 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 16:57:52 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:02:16 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:02:48 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:07:18 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:08:01 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:12:18 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:12:41 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:17:21 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3428.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:17:34 2010
    Restarting dead background process MMON
    MMON started with pid=11, OS id=5912
    Thu Jun 03 17:17:37 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:18:01 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:22:37 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:23:01 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:27:39 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:28:02 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:32:42 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:33:07 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:37:45 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_5912.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:38:40 2010
    Restarting dead background process MMON
    MMON started with pid=11, OS id=1660
    Thu Jun 03 17:38:43 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:39:17 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:42:54 2010
    The value (30) of MAXTRANS parameter ignored.
    kupprdp: master process DM00 started with pid=31, OS id=6116
    to execute - SYS.KUPM$MCP.MAIN('SYS_EXPORT_SCHEMA_01', 'SYSTEM', 'KUPC$C_1_20100603174259', 'KUPC$S_1_20100603174259', 0);
    Thu Jun 03 17:43:38 2010
    The value (30) of MAXTRANS parameter ignored.
    kupprdp: master process DM00 started with pid=32, OS id=2792
    to execute - SYS.KUPM$MCP.MAIN('SYS_EXPORT_SCHEMA_01', 'SYSTEM', 'KUPC$C_1_20100603174338', 'KUPC$S_1_20100603174338', 0);
    Thu Jun 03 17:43:44 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:44:06 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:44:47 2010
    The value (30) of MAXTRANS parameter ignored.
    kupprdp: master process DM00 started with pid=33, OS id=3492
    to execute - SYS.KUPM$MCP.MAIN('SYS_EXPORT_SCHEMA_01', 'SYSTEM', 'KUPC$C_1_20100603174448', 'KUPC$S_1_20100603174448', 0);
    kupprdp: worker process DW01 started with worker id=1, pid=34, OS id=748
    to execute - SYS.KUPW$WORKER.MAIN('SYS_EXPORT_SCHEMA_01', 'SYSTEM');
    Thu Jun 03 17:45:28 2010
    Memory Notification: Library Cache Object loaded into SGA
    Heap size 5684K exceeds notification threshold (2048K)
    KGL object name :SELECT /*+rule*/ SYS_XMLGEN(VALUE(KU$), XMLFORMAT.createFormat2('TABLE_T', '7')), KU$.OBJ_NUM ,KU$.ANC_OBJ.NAME ,KU$.ANC_OBJ.OWNER_NAME ,KU$.ANC_OBJ.TYPE_NAME ,KU$.BASE_OBJ.NAME ,KU$.BASE_OBJ.OWNER_NAME ,KU$.BASE_OBJ.TYPE_NAME ,KU$.SPARE1 ,KU$.XMLSCHEMACOLS ,KU$.SCHEMA_OBJ.NAME ,KU$.SCHEMA_OBJ.NAME ,'TABLE' ,KU$.PROPERTY ,KU$.SCHEMA_OBJ.OWNER_NAME ,KU$.TS_NAME ,KU$.TRIGFLAG FROM SYS.KU$_FHTABLE_VIEW KU$ WHERE NOT (BITAND (KU$.PROPERTY,8192)=8192) AND NOT BITAND(KU$.SCHEMA_OBJ.FLAGS,128)!=0 AND KU$.OBJ_NU
    Thu Jun 03 17:45:28 2010
    Memory Notification: Library Cache Object loaded into SGA
    Heap size 5681K exceeds notification threshold (2048K)
    Details in trace file c:\oraclexe\app\oracle\admin\xe\bdump\xe_dw01_748.trc
    KGL object name :SELECT /*+rule*/ SYS_XMLGEN(VALUE(KU$), XMLFORMAT.createFormat2('TABLE_T', '7')), KU$.OBJ_NUM ,KU$.ANC_OBJ.NAME ,KU$.ANC_OBJ.OWNER_NAME ,KU$.ANC_OBJ.TYPE_NAME ,KU$.BASE_OBJ.NAME ,KU$.BASE_OBJ.OWNER_NAME ,KU$.BASE_OBJ.TYPE_NAME ,KU$.SPARE1 ,KU$.XMLSCHEMACOLS ,KU$.SCHEMA_OBJ.NAME ,KU$.SCHEMA_OBJ.NAME ,'TABLE' ,KU$.PROPERTY ,KU$.SCHEMA_OBJ.OWNER_NAME ,KU$.TS_NAME ,KU$.TRIGFLAG FROM SYS.KU$_FHTABLE_VIEW KU$ WHERE NOT (BITAND (KU$.PROPERTY,8192)=8192) AND NOT BITAND(KU$.SCHEMA_OBJ.FLAGS,128)!=0 AND KU$.OBJ_NU
    Thu Jun 03 17:48:47 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:49:17 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:53:49 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Thu Jun 03 17:54:28 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_1660.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Dump file c:\oraclexe\app\oracle\admin\xe\bdump\alert_xe.log
    Fri Jun 04 07:46:55 2010
    ORACLE V10.2.0.1.0 - Production vsnsta=0
    vsnsql=14 vsnxtr=3
    Windows XP Version V5.1 Service Pack 3
    CPU : 2 - type 586, 1 Physical Cores
    Process Affinity : 0x00000000
    Memory (Avail/Total): Ph:1653M/2047M, Ph+PgF:4706M/4958M, VA:1944M/2047M
    Fri Jun 04 07:46:55 2010
    Starting ORACLE instance (normal)
    Fri Jun 04 07:47:06 2010
    LICENSE_MAX_SESSION = 100
    LICENSE_SESSIONS_WARNING = 80
    Picked latch-free SCN scheme 2
    Using LOG_ARCHIVE_DEST_10 parameter default value as USE_DB_RECOVERY_FILE_DEST
    Autotune of undo retention is turned on.
    IMODE=BR
    ILAT =33
    LICENSE_MAX_USERS = 0
    SYS auditing is disabled
    ksdpec: called for event 13740 prior to event group initialization
    Starting up ORACLE RDBMS Version: 10.2.0.1.0.
    System parameters with non-default values:
    processes = 200
    sessions = 300
    license_max_sessions = 100
    license_sessions_warning = 80
    sga_max_size = 838860800
    __shared_pool_size = 260046848
    shared_pool_size = 209715200
    __large_pool_size = 25165824
    __java_pool_size = 4194304
    __streams_pool_size = 8388608
    spfile = C:\ORACLEXE\APP\ORACLE\PRODUCT\10.2.0\SERVER\DBS\SPFILEXE.ORA
    sga_target = 734003200
    control_files = C:\ORACLEXE\ORADATA\XE\CONTROL.DBF
    __db_cache_size = 432013312
    compatible = 10.2.0.1.0
    db_recovery_file_dest = D:\
    db_recovery_file_dest_size= 5368709120
    undo_management = AUTO
    undo_tablespace = UNDO
    remote_login_passwordfile= EXCLUSIVE
    dispatchers = (PROTOCOL=TCP) (SERVICE=XEXDB)
    shared_servers = 10
    job_queue_processes = 1000
    audit_file_dest = C:\ORACLEXE\APP\ORACLE\ADMIN\XE\ADUMP
    background_dump_dest = C:\ORACLEXE\APP\ORACLE\ADMIN\XE\BDUMP
    user_dump_dest = C:\ORACLEXE\APP\ORACLE\ADMIN\XE\UDUMP
    core_dump_dest = C:\ORACLEXE\APP\ORACLE\ADMIN\XE\CDUMP
    db_name = XE
    open_cursors = 300
    os_authent_prefix =
    pga_aggregate_target = 209715200
    PMON started with pid=2, OS id=3044
    MMAN started with pid=4, OS id=3052
    DBW0 started with pid=5, OS id=3196
    LGWR started with pid=6, OS id=3200
    CKPT started with pid=7, OS id=3204
    SMON started with pid=8, OS id=3208
    RECO started with pid=9, OS id=3212
    CJQ0 started with pid=10, OS id=3216
    MMON started with pid=11, OS id=3220
    MMNL started with pid=12, OS id=3224
    Fri Jun 04 07:47:31 2010
    starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
    starting up 10 shared server(s) ...
    Oracle Data Guard is not available in this edition of Oracle.
    PSP0 started with pid=3, OS id=3048
    Fri Jun 04 07:47:41 2010
    alter database mount exclusive
    Fri Jun 04 07:47:54 2010
    Setting recovery target incarnation to 2
    Fri Jun 04 07:47:56 2010
    Successful mount of redo thread 1, with mount id 2601933156
    Fri Jun 04 07:47:56 2010
    Database mounted in Exclusive Mode
    Completed: alter database mount exclusive
    Fri Jun 04 07:47:57 2010
    alter database open
    Fri Jun 04 07:48:00 2010
    Beginning crash recovery of 1 threads
    Fri Jun 04 07:48:01 2010
    Started redo scan
    Fri Jun 04 07:48:03 2010
    Completed redo scan
    16441 redo blocks read, 442 data blocks need recovery
    Fri Jun 04 07:48:04 2010
    Started redo application at
    Thread 1: logseq 1575, block 48102
    Fri Jun 04 07:48:05 2010
    Recovery of Online Redo Log: Thread 1 Group 1 Seq 1575 Reading mem 0
    Mem# 0 errs 0: D:\XE\ONLINELOG\O1_MF_1_4CT6N7TC_.LOG
    Fri Jun 04 07:48:07 2010
    Completed redo application
    Fri Jun 04 07:48:07 2010
    Completed crash recovery at
    Thread 1: logseq 1575, block 64543, scn 27413940
    442 data blocks read, 442 data blocks written, 16441 redo blocks read
    Fri Jun 04 07:48:09 2010
    LGWR: STARTING ARCH PROCESSES
    ARC0 started with pid=25, OS id=3288
    ARC1 started with pid=26, OS id=3292
    Fri Jun 04 07:48:10 2010
    ARC0: Archival started
    ARC1: Archival started
    LGWR: STARTING ARCH PROCESSES COMPLETE
    Thread 1 advanced to log sequence 1576
    Thread 1 opened at log sequence 1576
    Current log# 3 seq# 1576 mem# 0: D:\XE\ONLINELOG\O1_MF_3_4CT6N1SD_.LOG
    Successful open of redo thread 1
    Fri Jun 04 07:48:13 2010
    ARC0: STARTING ARCH PROCESSES
    Fri Jun 04 07:48:13 2010
    ARC1: Becoming the 'no FAL' ARCH
    Fri Jun 04 07:48:13 2010
    ARC1: Becoming the 'no SRL' ARCH
    Fri Jun 04 07:48:13 2010
    ARC2: Archival started
    ARC0: STARTING ARCH PROCESSES COMPLETE
    ARC0: Becoming the heartbeat ARCH
    Fri Jun 04 07:48:13 2010
    SMON: enabling cache recovery
    ARC2 started with pid=27, OS id=3580
    Fri Jun 04 07:48:17 2010
    db_recovery_file_dest_size of 5120 MB is 49.00% used. This is a
    user-specified limit on the amount of space that will be used by this
    database for recovery-related files, and does not reflect the amount of
    space available in the underlying filesystem or ASM diskgroup.
    Fri Jun 04 07:48:31 2010
    Successfully onlined Undo Tablespace 1.
    Fri Jun 04 07:48:31 2010
    SMON: enabling tx recovery
    Fri Jun 04 07:48:31 2010
    Database Characterset is AL32UTF8
    replication_dependency_tracking turned off (no async multimaster replication found)
    Starting background process QMNC
    QMNC started with pid=28, OS id=2412
    Fri Jun 04 07:48:51 2010
    Completed: alter database open
    Fri Jun 04 07:49:22 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Fri Jun 04 07:49:32 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Fri Jun 04 07:49:52 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Fri Jun 04 07:49:57 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Fri Jun 04 07:54:10 2010
    Shutting down archive processes
    Fri Jun 04 07:54:15 2010
    ARCH shutting down
    ARC2: Archival stopped
    Fri Jun 04 07:54:53 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Fri Jun 04 07:55:08 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Fri Jun 04 07:56:25 2010
    Starting control autobackup
    Fri Jun 04 07:56:27 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\udump\xe_ora_488.trc:
    Fri Jun 04 07:56:28 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\udump\xe_ora_488.trc:
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_21
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_20
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_17
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_16
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_14
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_12
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_09
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_07
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_06
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\AUTOBACKUP\2009_04_03
    ORA-27093: 无法删除目录
    Fri Jun 04 07:56:29 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\udump\xe_ora_488.trc:
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_21
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_20
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_17
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_16
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_14
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_12
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_09
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_07
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_06
    ORA-27093: 无法删除目录
    ORA-17624: 无法删除目录 D:\XE\BACKUPSET\2009_04_03
    ORA-27093: 无法删除目录
    Control autobackup written to DISK device
    handle 'D:\XE\AUTOBACKUP\2010_06_04\O1_MF_S_720777385_60JJ9BNZ_.BKP'
    Fri Jun 04 07:56:38 2010
    Thread 1 advanced to log sequence 1577
    Current log# 1 seq# 1577 mem# 0: D:\XE\ONLINELOG\O1_MF_1_4CT6N7TC_.LOG
    Fri Jun 04 07:56:56 2010
    Thread 1 cannot allocate new log, sequence 1578
    Checkpoint not complete
    Current log# 1 seq# 1577 mem# 0: D:\XE\ONLINELOG\O1_MF_1_4CT6N7TC_.LOG
    Thread 1 advanced to log sequence 1578
    Current log# 3 seq# 1578 mem# 0: D:\XE\ONLINELOG\O1_MF_3_4CT6N1SD_.LOG
    Fri Jun 04 07:57:04 2010
    Memory Notification: Library Cache Object loaded into SGA
    Heap size 2208K exceeds notification threshold (2048K)
    KGL object name :XDB.XDbD/PLZ01TcHgNAgAIIegtw==
    Fri Jun 04 07:59:54 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Fri Jun 04 07:59:58 2010
    Errors in file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_3220.trc:
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []

    Hi Gurus,
    there's a error ora-00600 in the big trc files as below, this is only part of this file, this file is more than 45mb in size:
    xe_mmon_4424.trc
    Dump file c:\oraclexe\app\oracle\admin\xe\bdump\xe_mmon_4424.trc
    Fri Jun 04 17:03:22 2010
    ORACLE V10.2.0.1.0 - Production vsnsta=0
    vsnsql=14 vsnxtr=3
    Oracle Database 10g Express Edition Release 10.2.0.1.0 - Production
    Windows XP Version V5.1 Service Pack 3
    CPU : 2 - type 586, 1 Physical Cores
    Process Affinity : 0x00000000
    Memory (Avail/Total): Ph:992M/2047M, Ph+PgF:3422M/4958M, VA:1011M/2047M
    Instance name: xe
    Redo thread mounted by this instance: 1
    Oracle process number: 11
    Windows thread id: 4424, image: ORACLE.EXE (MMON)
    *** SERVICE NAME:(SYS$BACKGROUND) 2010-06-04 17:03:22.265
    *** SESSION ID:(284.23) 2010-06-04 17:03:22.265
    *** 2010-06-04 17:03:22.265
    ksedmp: internal or fatal error
    ORA-00600: internal error code, arguments: [kjhn_post_ha_alert0-862], [], [], [], [], [], [], []
    Current SQL statement for this session:
    BEGIN :success := dbms_ha_alerts_prvt.check_ha_resources; END;
    ----- PL/SQL Call Stack -----
    object line object
    handle number name
    41982E80 418 package body SYS.DBMS_HA_ALERTS_PRVT
    41982E80 552 package body SYS.DBMS_HA_ALERTS_PRVT
    41982E80 305 package body SYS.DBMS_HA_ALERTS_PRVT
    419501A0 1 anonymous block
    ----- Call Stack Trace -----
    calling call entry argument values in hex
    location type point (? means dubious value)
    ksedst+38           CALLrel  ksedst1+0 0 1
    ksedmp+898          CALLrel  ksedst+0 0
    ksfdmp+14           CALLrel  ksedmp+0 3
    _kgerinv+140         CALLreg  00000000             8EF0A38 3
    kgeasnmierr+19      CALLrel  kgerinv+0 8EF0A38 6610020 3672F70 0
    6538808
    kjhnpost_ha_alert CALLrel _kgeasnmierr+0       8EF0A38 6610020 3672F70 0
    0+2909
    __PGOSF57__kjhn_pos CALLrel kjhnpost_ha_alert 88 B21C4D0 B21C4D8 B21C4E0
    t_ha_alert_plsql+43 0+0 B21C4E8 B21C4F0 B21C4F8
    8 B21C500 B21C50C 0 FFFFFFFF 0
    0 0 6
    _spefcmpa+415        CALLreg  00000000            
    spefmccallstd+147   CALLrel  spefcmpa+0 65395B8 16 B21C5AC 653906C 0
    pextproc+58         CALLrel  spefmccallstd+0 6539874 6539760 6539628
    65395B8 0
    __PGOSF302__peftrus CALLrel _pextproc+0         
    ted+115
    _psdexsp+192         CALLreg  00000000             6539874
    _rpiswu2+426         CALLreg  00000000             6539510
    psdextp+567         CALLrel  rpiswu2+0 41543288 0 65394F0 2 6539528
    0 65394D0 0 2CD9E68 0 6539510
    0
    _pefccal+452         CALLreg  00000000            
    pefcal+174          CALLrel  pefccal+0 6539874
    pevmFCAL+128 CALLrel _pefcal+0           
    pfrinstrFCAL+55 CALLrel pevmFCAL+0 AF74F48 3DFB92B8
    pfrrunno_tool+56 CALL??? 00000000 AF74F48 3DFBB728 AF74F84
    pfrrun+781          CALLrel  pfrrun_no_tool+0 AF74F48 3DFBB28C AF74F84
    plsqlrun+738 CALLrel _pfrrun+0            AF74F48
    peicnt+247          CALLrel  plsql_run+0 AF74F48 1 0
    kkxexe+413          CALLrel  peicnt+0
    opiexe+5529         CALLrel  kkxexe+0 AF7737C
    kpoal8+2165         CALLrel  opiexe+0 49 3 653A4FC
    _opiodr+1099         CALLreg  00000000             5E 0 653CBAC
    kpoodr+483          CALLrel  opiodr+0
    _xupirtrc+1434       CALLreg  00000000             67384BC 5E 653CBAC 0 653CCBC
    upirtrc+61          CALLrel  xupirtrc+0 67384BC 5E 653CBAC 653CCBC
    653D990 60FEF8B8 653E194
    6736CD8 1 0 0
    kpurcsc+100         CALLrel  upirtrc+0 67384BC 5E 653CBAC 653CCBC
    653D990 60FEF8B8 653E194
    6736CD8 1 0 0
    kpuexecv8+2815      CALLrel  kpurcsc+0
    kpuexec+2106        CALLrel  kpuexecv8+0 673AE10 6736C4C 6736CD8 0 0
    653EDE8
    OCIStmtExecute+29   CALLrel  kpuexec+0 673AE10 6736C4C 673AEC4 1 0 0
    0 0 0
    kjhnmmon_action+5 CALLrel _OCIStmtExecute+0    673AE10 6736C4C 673AEC4 1 0 0
    26 0 0
    kjhncheck_ha_reso CALLrel kjhnmmon_action+0 653EFCC 3E
    urces+140
    kebmronce_dispatc CALL??? 00000000
    her+630
    kebmronce_execute CALLrel kebmronce_dispatc
    +12 her+0
    _ksbcti+788          CALLreg  00000000             0 0
    ksbabs+659          CALLrel  ksbcti+0
    kebmmmon_main+386 CALLrel _ksbabs+0            3C5DCB8
    _ksbrdp+747          CALLreg  00000000             3C5DCB8
    opirip+674          CALLrel  ksbrdp+0
    opidrv+857          CALLrel  opirip+0 32 4 653FEBC
    sou2o+45            CALLrel  opidrv+0 32 4 653FEBC
    opimaireal+227 CALLrel _sou2o+0             653FEB0 32 4 653FEBC
    opimai+92           CALLrel  opimai_real+0 3 653FEE8
    BackgroundThreadSt  CALLrel  opimai+0
    art@4+422
    7C80B726 CALLreg 00000000
    --------------------- Binary Stack Dump ---------------------
    ========== FRAME [1] (_ksedst+38 -> _ksedst1+0) ==========
    Dump of memory from 0x065386DC to 0x065386EC
    65386D0 065386EC [..S.]
    65386E0 0040467B 00000000 00000001 [{F@.........]
    ========== FRAME [2] (_ksedmp+898 -> _ksedst+0) ==========
    Dump of memory from 0x065386EC to 0x065387AC
    65386E0 065387AC [..S.]
    65386F0 00403073 00000000 53532E49 20464658 [[email protected] ]
    6538700 54204D41 0000525A 00000000 08EF0EC0 [AM TZR..........]
    6538710 6072D95A 08EF0EC5 03672F70 00000017 [Z.r`....p/g.....]
    6538720 00000000 00000000 00000000 00000000 [................]
    Repeat 1 times
    6538740 00000000 00000000 00000000 00000017 [................]
    6538750 08EF0B3C 08EF0B34 03672F70 08F017F0 [<...4...p/g.....]
    6538760 603AA0D3 065387A8 00000001 00000000 [..:`..S.........]
    6538770 00000000 00000000 00000001 00000000 [................]
    6538780 00000000 08EF0A38 06610020 031E1D20 [....8... .a. ...]
    6538790 00000000 065386F8 08EF0A38 06538D38 [......S.8...8.S.]
    65387A0 0265187C 031C8860 FFFFFFFF [|.e.`.......]
    ========== FRAME [3] (_ksfdmp+14 -> _ksedmp+0) ==========
    and the file is keeping increasing, though I have deleted a lot of this, but:
    as I marked:
    time size
    15:23 pm 795mb
    16:55 pm 959mb
    17:01 pm 970mb
    17:19 pm 990mb
    Any solution for that?
    Thanks!!

  • Am I really the winner of a very large sum of mone...

    Yesterday, I received a post on Skype, from a christopher.foster16 who wanted to make contact with me, but also informed me that I was the winner of a very large sum of money in Great British Pounds. He gave security ? numbers and a name I must contact to have the money sent to my address which I was asked to give in an email to Mr Petersen whitehead skype cash [email protected]
    The prize amount is very tempting ofcourse, but my son said it has to be checked out as he thinks it is a definite scam.  what can I do about this?...anyone know Please. I am 83 years old, and only use skype to contact my relatives and friends who are overseas.
    Thankx.
    Post transferred from blog article;
    subject/title edited accordingly.

    correct, cofirming that is indeed a very well known scam.  you can also report the user/sender to [email protected]
    IF YOU FOUND OUR POST USEFUL THEN PLEASE GIVE "KUDOS". IF IT HELPED TO FIX YOUR ISSUE PLEASE MARK IT AS A "SOLUTION" TO HELP OTHERS. THANKS!
    ALTERNATIVE SKYPE DOWNLOAD LINKS | HOW TO RECORD SKYPE VIDEO CALLS | HOW TO HANDLE SUSPICIOS CALLS AND MESSAGES
    SEE MORE TIPS, TRICKS, TUTORIALS AND UPDATES in
    | skypefordummies.blogspot.com | 

  • Best data Structor for dealing with very large CSV files

    hi im writeing an object that stores data from a very large CSV file. The idea been that you initlize the object with the CSV file, then it has lots of methods to make manipulating and working with the CSV file simpler. Operations like copy colum, eliminate rows, perform some equations on all values in a certain colum, etc. Also a method for prining back to a file.
    however the CSV files will probly be in the 10mb range maby larger so simply loading into an array isn't posable. as it produces a outofmemory error.
    does anyone have a data structor they could recomend that can store the large amounts of data require and are easly writeable. i've currently been useing a randomaccessfile but it is aquard to write to as well as needing an external file which would need to been cleaned up after the object is removed (something very hard to guarentee occurs).
    any suggestions would be greatly apprechiated.
    Message was edited by:
    ninjarob

    How much internal storage ("RAM") is in the computer where your program should run? I think I have 640 Mb in mine, and I can't believe loading 10 Mb of data would be prohibitive, not even if the size doubles when the data comes into Java variables.
    If the data size turns out to be prohibitive of loading into memory, how about a relational database?
    Another thing you may want to consider is more object-oriented (in the sense of domain-oriented) analysis and design. If the data is concerned with real-life things (persons, projects, monsters, whatever), row and column operations may be fine for now, but future requirements could easily make you prefer something else (for example, a requirement to sort projects by budget or monsters by proximity to the hero).

  • Keeping two very large datastores in sync

    I'm looking at options for keeping a very large (potentially 400GB) TimesTen (11.2.2.5) datastore in sync between a Production server and a [warm] Standby.
    Replication has been discounted because it doesn't support compressed tables, nor the types of table our closed-code application is creating (without non-null PKs)
    I've done some testing with smaller datastores to get indicative numbers, and a 7.4GB datastore (according to dssize) resulted in a 35GB backup set (using ttBackup -type fileIncrOrFull). Is that large increase in volume expected, and would it extrapolate up for a 400GB data store (2TB backup set??)?
    I've seen that there are Incremental backups, but to maintain our standby as warm, we'll be restoring these backups and from what I'd read & tested only a ttDestroy/ttRestore is possible, i.e. complete restore of the complete DSN each time, which is time consuming. Am I missing a smarter way of doing this?
    Other than building our application to keep the two datastores in sync, are there any other tricks we can use to efficiently keep the two datastores in sync?
    Random last question - I see "datastore" and "database" (and to an extent, "DSN") used apparently interchangeably - are they the same thing in TimesTen?
    Update: the 35GB compresses down with 7za to just over 2.2GB, but takes 5.5 hours to do so. If I take a standalone fileFull backup it is just 7.4GB on disk, and completes faster too.
    thanks,
    rmoff.
    Message was edited by: rmoff - add additional detail

    This must be an Exalytics system, right? I ask this because compressed tables are not licensed for use outside of an Exalytics system...
    As you note, currently replication is not possible in an Exalytics environment, but that is likely to change in the future and then it will definitely be the preferred mechanism for this. There is not really any other viable way to do this other than through the application.
    With regard to your specific questions:
    1.   A backup consists primarily of the most recent checkpoint file plus all log files/records that are newer than that file. So, to minimise the size of a full backup ensure
         that a checkpoint occurs (for example 'call ttCkpt' from a ttIsql session) immediately prior to starting the backup.
    2.   No, only complete restore is possible from an incremental backup set. Also note that due to the large amount of rollforward needed, restoring a large incremental backup set may take quite a long time. Backup and restore are not really intended for this purpose.
    3.   If you cannot use replication then some kind of application level sync is your only option.
    4.   Datastore and database mean the same thing - a physical TimesTen database. We prefer the term database nowadays; datastore is a legacy term. A DSN is a different thing (Data Source Name) and should not be used interchangeably with datastore/database. A DSN is a logical entity that defines the attributes for a database and how to connect to it. It is not the same as a database.
    Chris

  • Very large values from accelerome​ter

    Hello,
    I did a small impact test in the lab, and recorded the acceleration. But I got very large peak numbers, like 300g, which is unreasonable. My vi has been checked by many people, seems like no problem. Does anyone have the idea why I got such big numbers?
    Thanks.
    Yuan

    It would be helpful if you describe your hardware, DAQ , (signal-conditioner with setting) , sensor
    and please post a link to the spec-sheet of the sensor and (if you have) the individual sensitivity (cal) value(s) of your sensor. (in metric if possible )
    Now read the input voltage at a high sample rate (you can use one of the example vi's ) . If you have a pulse in the diagram hit edit-make current values default, save as a new vi and post that vi.
    Greetings from Germany
    Henrik
    LV since v3.1
    “ground” is a convenient fantasy
    '˙˙˙˙uıɐƃɐ lɐıp puɐ °06 ǝuoɥd ɹnoʎ uɹnʇ ǝsɐǝld 'ʎɹɐuıƃɐɯı sı pǝlɐıp ǝʌɐɥ noʎ ɹǝqɯnu ǝɥʇ'

  • Very Large Snapshot

    Hi
    We are using Hyper-V on Server 2012, we have 4 nodes in a cluster and the machines are kept on clustered storage.
    During some maintenance on one of our SQL servers I noticed that the VM had been snapshotted way back in January 2014.  This has led to some very large snapshots.
    The VM has 3 disks - an OS disk (running Server 2008R2 with SQL Server 2008R2), a disk for the DBs and a disk for the SQL transaction logs.  ALL disks have large (70GB+) snapshots against them.
    I don't know why this machine was snapshotted - I don't remember ever doing it myself (and certainly from experience using VM Ware Snapshots in the past steer well clear of these things).
    My main concern is not only are they large snapshots but the VM is a busy SQL server.  We migrated from VM Ware a year ago and I have never used Hyper-V snapshots, a bit of research shows up what I would expect (don't use for production unless necessary,
    don't use for DB/AD/Exchange servers etc).
    So my question to the community is what course of action should I take?  Delete the snapshots and hope for the best?
    Our DBs are backed up every 15mins and the OS is replicated offsite so if it went wrong then recovery is possible (but not painless!)
    Does anyone have any experience of such a situation?
    Any advice greatly appreciated!

    I have very little information in the clients setup outside its server 2012 and he has an unknown version of SQL server
    VM has 3 what, VHD or bare disks?
    I need more facts
    No, you don't, the problem was quite clearly stated in the original post. He has some large snapshots, doesn't know who took them or why, and wants to know if it is safe to delete them. A very simple question and one which has already been answered by Sam.
    It is obvious from your responses that you simply don't understand what this thread is all about. Providing you with more information is pointless. It would just be more information that you don't understand.
    I assumed he wanted new ones
    Corsair Carbide 300R with TX850V2
    Asus M5A99FX PRO R2.0 CFX/SLI
    AMD Phenom II 965 C3 Black Edition @ 4.0 GHz
    G.SKILL RipjawsX DDR3-2133 8 GB
    EVGA GTX 660 Ti FTW Signature 2 (GK104 Kepler)
    Asus PA238QR IPS LED HDMI DP 1080p
    ST2000DM001 & Windows 8.1 Professional x64
    Microsoft Wireless Desktop 2000 & Wacom Bamboo CHT470M
    Place your rig specifics into your signature like I have, makes it 100x easier to understand!
    Hardcore Games Legendary is the Only Way to Play!

  • Maximmum value of a sequnec and viewing very large numbers

    Hi,
    We have a table which has a column populated using a sequence - it is maintained by code in owb but growing at larger rate than expect
    e.g sequence starts at say 10000000000, would expect next row to b created with 10000000001 but large gap in between.
    Have a separate ticket raised with oracel for this as mainatined via dimesnion opertaor code.
    However, have couple of questions reagrding sequences.
    1) How large can they be
    2) If we try query numbers which over 14 digits long starts to show e in the tool we are using (pl/sql developer) Is there a way to ensure we can see the whole number (sqlplus?) and
    I'm assuming say if had number column which very large say 18 digits long joined to another table 18 digits long then no issues?
    Thanks

    1. you really should read docs.
    e.g sequence starts at say 10000000000, would expect next row to b created with 10000000001 but large gap in between.Not necessarily. Oracle Sequence guarantees uniqueness, not continuity.
    There can be gaps by design. Especially if CACHE of the sequence is big.
    Have a separate ticket raised with oracel for this as mainatined via dimesnion opertaor code.you really should read docs before creating tickets.
    However, have couple of questions reagrding sequences.
    1) How large can they be-As large as NUMBER datatype can be - 38 digits.-
    I am wrong. Doc says - 28 digits.
    2) If we try query numbers which over 14 digits long starts to show e in the tool we are using (pl/sql developer) Is there a way to ensure we can see the whole number (sqlplus?) What is printed on screen is a matter of formatting. You can choose formatting you like in SQLPlus and SQLDeveloper
    I'm assuming say if had number column which very large say 18 digits long joined to another table 18 digits long then no issues?No issues.
    Edited by: Mark Malakanov (user11181920) on Apr 10, 2013 2:06 PM

Maybe you are looking for