When loading large number of images, images begin to fail to load in IE11 with DOM: 7009 error (unable to decode) in console

When loading many images on a single page in IE (reproduced in IE11), some of them begin to fail to load, and have something similar to the following warning in the console:
DOM7009: Unable to decode image at URL: '[some unique url]'.
When I look at the network traffic, there does appear to valid responses received for each of these images from the server. It's not always the same images each time. If I use the dev tools to force the image to reload (example: I update the url to include some some extraneous url parameter "&test=1"), it loads/renders normally without error. I've reproduced this behavior with different types of images (jpegs/pngs; example png included below). Apologies for the cross-posting. I wasn't sure what the particular etiquette was for that - I've also posted this issue on StackOverflow and MS Connect under the same title. If one of those is a more appropriate place for this matter, please let me know and I'll close this issue.

yup, i have encountered this problem as well. I'm loading an array of image frames for a movie, and i'm using preloadJS (javascript preloading library) and not displaying the images upfront. 

Similar Messages

  • I have accidentally deleted a large number of develped images in Lightroom before I did a backup. I reimported the original raw files back into Lightroom hoping the develop settings would be re-established but no luck. Notice system mau have done an auto-

    Question?
    I have accidentally deleted a large number of develped images in Lightroom before I did a backup. I reimported the original raw files back into Lightroom hoping the develop settings would be re-established but no luck. Notice system mau have done an auto-backup as have an lrcat-journal file. Can I use this to restore my develop settings. I also have jpgs generated from all the deleted images.

    Hello,
    if you have a backup of your catalog you can do the following:
    1. Backup your catalog first
    2. Restore your backup catalog to some location
    3. Open your current catalog and select "files->import from another catalog".
    4. Select your backup catalog and your lost images. LR ask you if you want to overwrite the current settings or save them as a virtual copy.
    As an alternativ you can open your backup catalog, select the "lost" images and save the development settings as xmp sidecar fiels (using ctrl-s). Then open your current catalog, seletct the images and use "Metadata->Read Metadata from files".

  • When I try to install adobe photoshop 13 on my computer it comes up with the following error message " This installer does not support installation on a 64bit windows operating system" error code 6. How to I recify this? TIA

    When I try to install adobe photoshop 13 on my computer it comes up with the following error message " This installer does not support installation on a 64bit windows operating system" error code 6. How to I recify this? TIA

    You can try downloading the 64 bit version.  You can download the trial version of the software thru the page linked below and then use your current serial number to activate it.
    Be sure to follow the steps outlined in the Note: Very Important Instructions section on the download pages at this site and have cookies enabled in your browser or else the download will not work properly.
    Photoshop/Premiere Elements 13: http://prodesigntools.com/photoshop-elements-13-direct-download-links-premiere.html

  • Loading large number of coordinates

    Hello,
    We are trying to load the SDO_ORDINATES columnobject of the SDO_GEOMETRY column in a spatial feature table with a large number of coordinates (4844). This is not the largest linestring we will need to build. We have defined a procedure as:
    CREATE OR REPLACE PROCEDURE
    INSERT_GPS(GEOM MDSYS.SDO_GEOMETRY) IS
    BEGIN
    INSERT INTO GPS(long_lat_elv) VALUES (GEOM);
    COMMIT;
    END;
    and are loading the SDO_GEOMETRY column as follows:
    DECLARE
    myGEOM MDSYS.SDO_GEOMETRY := MDSYS.SDO_GEOMETRY(3002,NULL,NULL,MDSYS.SDO_ELEM_INFO_ARRAY(1,2,1),
    MDSYS.SDO_ORDINATE_ARRAY(
    -86.929501,35.751834,32.907,
    -86.929434,35.751906,32.913,
    87.270367,35.447903,.854,
    87.270273,35.447956,.86
    BEGIN
    dave_insert(myGEOM);
    END;
    This generates the following error:
    PLS-00123: Program too large.
    How can we load this many coordinates, using this workflow? We are using 8.1.6.
    Thanks
    Dave
    null

    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Tilen Skraba ([email protected]):
    Why don't you use SQL*Loader to load the data? We have geometries which have more data than yours and sqlloader didn't have any problems.
    I believe PLS-00123 means that your package is too big.
    What else do you have in your package?
    Try and break the package in more smaller ones.
    Tilen<HR></BLOCKQUOTE>
    Telim,
    We are trying to use the LRS package and the measure value, which is part of the coordinate (x,y,m) has to be generated from a stored procedure for each x,y pair.
    Thanks
    Dave
    null

  • MBPro crashes when transfering large number of files

    i have a MBPro 2008 and recently when i transfer a large number of video files [over 5GB] my computer will suddendly crash in the middle of the transfer.
    i'm usually not using it while its transfering. This is a very recent problem.
    Sincerly,
    stef./.\

    How large is you HD and how much space do you have left?
    Try the following:
    Disconnect all devices from the computer then do the following:
    Boot up from your Snow Leopard Install DVD while holding down the "c" key.
    Select the language you wish to use, and then choose "Disk Utility" from the menu bar.
    Your HD should appear in a panel on the left hand side of the window which opens. Select it and then click on the "repair disk" button in the right hand section of the window.
    Once this process has been completed restart your computer and repair permissions directly from Disk Utility.
    If Disk Utility is unable to complete the repair, you will need to use a stronger 3rd party utility like DiskWarrior, Techtool PRO (not Deluxe) or Drive Genius.

  • Making a 'Movie' from large number of 'still images?'

    I have a 2011 15" MacBook Pro with Lion. I have QuickTime Player 10.1.
    I want to make a 'movie' showing our new company product rotating on its axis. The raw material would be still images, each taken from a slightly different angle. I thought QuickTime Pro would be the tool for this, but apparently it no longer exists? Can I accomplish the same thing in iMovie?
    Suggestions most welcome.
    Thank you!
    Chris in New Hampshire

    Thank you, AppleMan! I just wanted to make sure that I'm OK buying this because the App Store tells me this as a warning:
    Please note: Your QuickTime 7 Pro registration code only works with QuickTime 7. Be sure you have downloaded and installed QuickTime 7 Player before you enter your QuickTime 7 Pro registration code.
    So, since I have QuickTime Player 10.1 I wanted to confirm that this isn't a roadblock.

  • Not receiving email when sending large number of records using a FM?

    Hi..
    I am using the function module " SO_DOCUMENT_SEND_API1 " to send email....
    When a single record is there.. or around 5-6 records are there... email is coming successfully...
    But when there are more records..say around 100, the email is not coming... I checked SOST transaction and the status there is in red..and the error message is " Internal error: SO_OBJECT_MIME_GET Exception: 2 ".........
    What could be the reason behind this problem....... ??
    I have another problem... my output has over 60 fields, but in the email which i am receiving has only around 10 fields... how to solve this problem...??
    Plz help...

    Well... right now i am tryin to get only the first 2 fields.. but even in this case... i am not getting the email if around 15 records are there......
    I am using the code which is given below which i found in SDN only......In this code.. data is getting selected from EKPO... i tried changing the number of rows getting selected.. and in this case, the attachement is coming as desired... but when i use the same code for my prog.. i am not getting the mail.. even if there are only 10 records or so...
    *& Report  ZT062108   ALV Header                                    *
    REPORT  zt062108.
    TABLES: ekko.
    PARAMETERS: p_email   TYPE somlreci1-receiver
                                      DEFAULT '<give email here>'.
    TYPES: BEGIN OF t_ekpo,
      ebeln TYPE ekpo-ebeln,
      ebelp TYPE ekpo-ebelp,
      aedat TYPE ekpo-aedat,
      matnr TYPE ekpo-matnr,
    END OF t_ekpo.
    DATA: it_ekpo TYPE STANDARD TABLE OF t_ekpo INITIAL SIZE 0,
          wa_ekpo TYPE t_ekpo.
    TYPES: BEGIN OF t_charekpo,
      ebeln(10) TYPE c,
      ebelp(5)  TYPE c,
      aedat(8)  TYPE c,
      matnr(18) TYPE c,
    END OF t_charekpo.
    DATA: wa_charekpo TYPE t_charekpo.
    DATA:   it_message TYPE STANDARD TABLE OF solisti1 INITIAL SIZE 0
                    WITH HEADER LINE.
    DATA:   it_attach TYPE STANDARD TABLE OF solisti1 INITIAL SIZE 0
                    WITH HEADER LINE.
    DATA:   t_packing_list LIKE sopcklsti1 OCCURS 0 WITH HEADER LINE,
            t_contents LIKE solisti1 OCCURS 0 WITH HEADER LINE,
            t_receivers LIKE somlreci1 OCCURS 0 WITH HEADER LINE,
            t_attachment LIKE solisti1 OCCURS 0 WITH HEADER LINE,
            t_object_header LIKE solisti1 OCCURS 0 WITH HEADER LINE,
            w_cnt TYPE i,
            w_sent_all(1) TYPE c,
            w_doc_data LIKE sodocchgi1,
            gd_error    TYPE sy-subrc,
            gd_reciever TYPE sy-subrc.
    *START_OF_SELECTION
    START-OF-SELECTION.
    *   Retrieve sample data from table ekpo
      PERFORM data_retrieval.
    *   Populate table with detaisl to be entered into .xls file
      PERFORM build_xls_data_table.
    *END-OF-SELECTION
    END-OF-SELECTION.
    * Populate message body text
      perform populate_email_message_body.
    * Send file by email as .xls speadsheet
      PERFORM send_file_as_email_attachment
                                   tables it_message
                                          it_attach
                                    using p_email
                                          'Example .xls documnet attachment'
                                          'TXT'
                                          'filename'
                                 changing gd_error
                                          gd_reciever.
    *   Instructs mail send program for SAPCONNECT to send email(rsconn01)
      PERFORM initiate_mail_execute_program.
    *&      Form  DATA_RETRIEVAL
    *       Retrieve data form EKPO table and populate itab it_ekko
    FORM data_retrieval.
      SELECT ebeln ebelp aedat matnr
       UP TO 1000 ROWS
        FROM ekpo
        INTO TABLE it_ekpo.
    ENDFORM.                    " DATA_RETRIEVAL
    *&      Form  BUILD_XLS_DATA_TABLE
    *       Build data table for .xls document
    FORM build_xls_data_table.
    *  CONSTANTS: con_cret TYPE x VALUE '0D'.  "OK for non Unicode
    *             con_tab TYPE x VALUE '09'.   "OK for non Unicode
    *If you have Unicode check active in program attributes thnen you will
    *need to declare constants as follows
    *class cl_abap_char_utilities definition load.
    constants:
        con_tab  type c value cl_abap_char_utilities=>HORIZONTAL_TAB,
        con_cret type c value cl_abap_char_utilities=>CR_LF.
      CONCATENATE 'EBELN' 'EBELP' 'AEDAT' 'MATNR'
             INTO it_attach SEPARATED BY con_tab.
      CONCATENATE con_cret it_attach  INTO it_attach.
      APPEND  it_attach.
      LOOP AT it_ekpo INTO wa_charekpo.
        CONCATENATE wa_charekpo-ebeln wa_charekpo-ebelp
                    wa_charekpo-aedat wa_charekpo-matnr
               INTO it_attach SEPARATED BY con_tab.
        CONCATENATE con_cret it_attach  INTO it_attach.
        APPEND  it_attach.
      ENDLOOP.
    ENDFORM.                    " BUILD_XLS_DATA_TABLE
    *&      Form  SEND_FILE_AS_EMAIL_ATTACHMENT
    *       Send email
    FORM send_file_as_email_attachment tables pit_message
                                              pit_attach
                                        using p_email
                                              p_mtitle
                                              p_format
                                              p_filename
                                              p_attdescription
                                              p_sender_address
                                              p_sender_addres_type
                                     changing p_error
                                              p_reciever.
      DATA: ld_error    TYPE sy-subrc,
            ld_reciever TYPE sy-subrc,
            ld_mtitle LIKE sodocchgi1-obj_descr,
            ld_email LIKE  somlreci1-receiver,
            ld_format TYPE  so_obj_tp ,
            ld_attdescription TYPE  so_obj_nam ,
            ld_attfilename TYPE  so_obj_des ,
            ld_sender_address LIKE  soextreci1-receiver,
            ld_sender_address_type LIKE  soextreci1-adr_typ,
            ld_receiver LIKE  sy-subrc.
      ld_email   = p_email.
      ld_mtitle = p_mtitle.
      ld_format              = p_format.
      ld_attdescription      = p_attdescription.
      ld_attfilename         = p_filename.
      ld_sender_address      = p_sender_address.
      ld_sender_address_type = p_sender_addres_type.
    * Fill the document data.
      w_doc_data-doc_size = 1.
    * Populate the subject/generic message attributes
      w_doc_data-obj_langu = sy-langu.
      w_doc_data-obj_name  = 'SAPRPT'.
      w_doc_data-obj_descr = ld_mtitle .
      w_doc_data-sensitivty = 'F'.
    * Fill the document data and get size of attachment
      CLEAR w_doc_data.
      READ TABLE it_attach INDEX w_cnt.
      w_doc_data-doc_size =
         ( w_cnt - 1 ) * 255 + STRLEN( it_attach ).
      w_doc_data-obj_langu  = sy-langu.
      w_doc_data-obj_name   = 'SAPRPT'.
      w_doc_data-obj_descr  = ld_mtitle.
      w_doc_data-sensitivty = 'F'.
      CLEAR t_attachment.
      REFRESH t_attachment.
      t_attachment[] = pit_attach[].
    * Describe the body of the message
      CLEAR t_packing_list.
      REFRESH t_packing_list.
      t_packing_list-transf_bin = space.
      t_packing_list-head_start = 1.
      t_packing_list-head_num = 0.
      t_packing_list-body_start = 1.
      DESCRIBE TABLE it_message LINES t_packing_list-body_num.
      t_packing_list-doc_type = 'RAW'.
      APPEND t_packing_list.
    * Create attachment notification
      t_packing_list-transf_bin = 'X'.
      t_packing_list-head_start = 1.
      t_packing_list-head_num   = 1.
      t_packing_list-body_start = 1.
      DESCRIBE TABLE t_attachment LINES t_packing_list-body_num.
      t_packing_list-doc_type   =  ld_format.
      t_packing_list-obj_descr  =  ld_attdescription.
      t_packing_list-obj_name   =  ld_attfilename.
      t_packing_list-doc_size   =  t_packing_list-body_num * 255.
      APPEND t_packing_list.
    * Add the recipients email address
      CLEAR t_receivers.
      REFRESH t_receivers.
      t_receivers-receiver = ld_email.
      t_receivers-rec_type = 'U'.
      t_receivers-com_type = 'INT'.
      t_receivers-notif_del = 'X'.
      t_receivers-notif_ndel = 'X'.
      APPEND t_receivers.
      CALL FUNCTION 'SO_DOCUMENT_SEND_API1'
           EXPORTING
                document_data              = w_doc_data
                put_in_outbox              = 'X'
                sender_address             = ld_sender_address
                sender_address_type        = ld_sender_address_type
                commit_work                = 'X'
           IMPORTING
                sent_to_all                = w_sent_all
           TABLES
                packing_list               = t_packing_list
                contents_bin               = t_attachment
                contents_txt               = it_message
                receivers                  = t_receivers
           EXCEPTIONS
                too_many_receivers         = 1
                document_not_sent          = 2
                document_type_not_exist    = 3
                operation_no_authorization = 4
                parameter_error            = 5
                x_error                    = 6
                enqueue_error              = 7
                OTHERS                     = 8.
    * Populate zerror return code
      ld_error = sy-subrc.
    * Populate zreceiver return code
      LOOP AT t_receivers.
        ld_receiver = t_receivers-retrn_code.
      ENDLOOP.
    ENDFORM.
    *&      Form  INITIATE_MAIL_EXECUTE_PROGRAM
    *       Instructs mail send program for SAPCONNECT to send email.
    FORM initiate_mail_execute_program.
      WAIT UP TO 2 SECONDS.
      SUBMIT rsconn01 WITH mode = 'INT'
                    WITH output = 'X'
                    AND RETURN.
    ENDFORM.                    " INITIATE_MAIL_EXECUTE_PROGRAM
    *&      Form  POPULATE_EMAIL_MESSAGE_BODY
    *        Populate message body text
    form populate_email_message_body.
      REFRESH it_message.
      it_message = 'Please find attached a list test ekpo records'.
      APPEND it_message.
    endform.                    " POPULATE_EMAIL_MESSAGE_BODY

  • Managed ODP: InvalidCastException when returning large number

    I get an InvalidCastException "Specified cast is not valid" when trying to get a value bigger than Decimal.MaxValue as a scalar result from a select.
    From the stack trace and inspection of the Oracle.ManagedDataAccess.dll code I would expect OracleDataReader to cast to OracleDecimal (or System.Numerics.BigInteger) instead of System.Decimal for large numbers.
    Is this a bug?
    Stack trace:
       at Oracle.ManagedDataAccess.Client.OracleDataReader.GetDecimal(Int32 i)
       at Oracle.ManagedDataAccess.Client.OracleDataReader.GetValue(Int32 i)
       at Oracle.ManagedDataAccess.Client.OracleCommand.ExecuteScalar()
       at Cora.Data.DBProvider.Tests.CoraParameterTest.BigIntegerOracleTest() in D:\svnroot\tech\lib\Cora.Data.DBProvider\Cora.Data.DBProvider.Tests\CoraParameterTest.cs:line 98
    How to reproduce:
    - create a table NUNITTEST_PARAM_SIMPLE with a column BIG of type NUMBER(38)
    - run sample code
                using ( OracleConnection conn = new OracleConnection(ORACONN) )
                    conn.Open();
                    using ( OracleCommand cmd = conn.CreateCommand() )
                        cmd.CommandText = "insert into NUNITTEST_PARAM_SIMPLE (BIG) values (1234567890123456789012345678901234567)";
                        cmd.ExecuteNonQuery();
                    using ( OracleCommand cmd = conn.CreateCommand() )
                        cmd.CommandText = "select BIG from NUNITTEST_PARAM_SIMPLE";
                        object result = cmd.ExecuteScalar();          // Exception
                    conn.Close();

    There is no safe type mapping for ExecuteScalar. It's an uncommon use case.
    Even if there was, making the correct conversion implicitly would lead to more work for the application developer. Let's assume ODP.NET performs the "right" conversion. Now, the app needs to manipulate the data. What data type does it treat the returned value as: string or decimal? It could be either. That means twice as much code. One set deals with the data if a decimal was returned. The second deals with the data if a string is returned.
    It would be much easier to write code for one scenario, rather than two. That is why ODP.NET safe type mapping is an explicit conversion the developer makes.

  • Error "java.lang.OutOfMemoryError" When Returning Large Number of Docs

    In our SES implementation, we have a custom search interface that allows users to search for documents and then add them to a "shopping cart". Users add then to their shopping cart from search results, by adding docs one-by-one or an Add All option. Once they are done "shopping" they create a report.
    Here is the scenario...
    Users are saerching for documents and seeing that on page 1, there is 1 - 10 of about 300 results. They clicked Add All and want all 300 docs added to their cart.
    What we do under the covers is we execute another search and set the docs requested to 200. We get the array of docs, iterate over them and add the keys to a list. We found 200 docs at a time to be a safe number. However, there are still 100 docs that were not added to their cart and users want all 300 added. In other words, when they click Add All, they want to add all docs, 300, 500, 5000, etc.
    I set the "Maximum Number of Results" to 500 and found that I can safely add up to ~ 350 docs at one time. However, going past this point throws the following error:
    [SOAPException: faultCode=SOAP-ENV:Server; msg= [java.lang.OutOfMemoryError]]
         at oracle.search.query.webservice.client.OracleSearchService.makeSOAPCallRPC(OracleSearchService.java:941)
         at oracle.search.query.webservice.client.OracleSearchService.doOracleSearch(OracleSearchService.java:469)
    After this error is thrown, SES was unable to recover and searching would not work anymore, even return 10 docs at a time. We had to restart SES to resolve the issue.
    1. What is throwing this error? Is it the amount of XML being returned?
    2. What is the maximum number of results we can get back at any one time? Is it based on the amount of data being returned or the number of attributes?
    We are running 10.1.8 with plans to upgrade soon.
    Thanks in advance.

    I know it may be hard to throw away all this code, but consider using the jakarta fileupload component.
    I think it would simplify your code down to
    // Create a factory for disk-based file items
    FileItemFactory factory = new DiskFileItemFactory();
    // Create a new file upload handler
    ServletFileUpload upload = new ServletFileUpload(factory);
    // Parse the request
    List /* FileItem */ items = upload.parseRequest(request);
    // Process the uploaded items
    Iterator iter = items.iterator();
    while (iter.hasNext()) {
        FileItem item = (FileItem) iter.next();
        if (item.isFormField()) {
            processFormField(item);
        } else {
            // item is a file.  write it
            File saveFolder = application.getRealPath("/file");          
            File uploadedFile = new File(saveFolder, item.getName());
            item.write(uploadedFile);
    }Most of this code was hijacked from http://jakarta.apache.org/commons/fileupload/using.html
    Check it out. It will solve your memory problem by writing the file to disk temporarily if necessary.
    Cheers,
    evnafets

  • ActiveX Printer Error when print large number of page

    Hi,
    I have this problem when developing the report on my project.
    I developed the Crystal Report embedded in Visual Studio 2008 to be displayed in web form.
    I'm using .NET and the operating system of the development computer is Windows 7 32 bit.
    The problem is, I have a quite complex report displaying the statement for the customer which is result in 209 pages.
    When the report has successfully loaded, I simply click the print icon and it comes up with the error "A communication error occurred. Printing will be stopped"
    But, after that...I tried something different, when the report has successfully loaded, I went through the report by moving for every 30 pages until it reaches the last page and click the print icon. and it prints without any errors!
    Anybody knows what caused this?
    By the way, I'm using the ActiveX printing mode for the Crystal Report.
    Thanks
    -Marry-

    I'm thinking there is something not processed in the report so forcing the report to process everything up to the last page before printing may help. Try the following code before sending the report to the viewer:
    Dim prc As New CrystalDecisions.Shared.ReportPageRequestContext
    rptDoc.FormatEngine.GetLastPageNumber(prc)
    This may slow down the time it takes for the report to come up as the whole report will need to be processed, formatted, etc.
    Also, make sure you have SP 1 for CR 10.5:
    https://smpdl.sap-ag.de/~sapidp/012002523100009351512008E/crbasic2008sp1.exe
    Ludek
    Follow us on Twitter http://twitter.com/SAPCRNetSup
    Got Enhancement ideas? Try the [SAP Idea Place|https://ideas.sap.com/community/products_and_solutions/crystalreports]

  • How to update the large amount of thumbnail images in background process

    in my application i want to load large number of thumbnail images(each images set as icon for separate jradiobutton) into panel. This panel is in left side of JSplitPane.
    when application is starting, now I’m loading the thumbnail buttons without image(to run the application quickly).
    now i want to update the buttons with real thumbnail images in background. i tried with thread and SwingUtilities.invokeLater ,but it stuck the application until updating finish.
    Note:im using java1.4 (not in the possession to use other versions)
    Can anybody give suggestion?
    SwingUtilities.invokeLater(new Runnable() {
    public void run() {
         for (int j = 0; j < imgPagesV.size(); j++) {
         try {
              ImageIcon icon = new ImageIcon((BufferedImage)thumImagesV.get(j)); //thumImagesV vector have thumnail bufferedimages 
              ((JRadioButton)thumButtonPanel.getComponent(j)).setIcon(icon);
              updateUI();
         } catch (Exception e) {
              // TODO Auto-generated catch block
              e.printStackTrace();
    });

    thax for your reply ..
    even i call new smiple thread to load the images it freeze the main gui. is any wrong in my way ?
    Thread Class
    import javax.media.jai.NullOpImage;
    import javax.media.jai.OpImage;
    import com.sun.media.jai.codec.ImageCodec;
    import com.sun.media.jai.codec.ImageDecoder;
    import com.sun.media.jai.codec.TIFFDecodeParam;
    public class TestThread extends Thread {
         File tiffImg;
         TestThread(File img) {
              this.tiffImg = img;
          * extracting form tiff images and creates thumbnail then add into vector
         public void run() {
              ImageDecoder dec = null;
              Vector thumImagesV = new Vector();
              TIFFDecodeParam param = new TIFFDecodeParam();
              param.setDecodePaletteAsShorts(true);
              param.setJPEGDecompressYCbCrToRGB(true);
              try {
                   dec = ImageCodec.createImageDecoder("tiff", tiffImg, param);
                   int start, end = 0;
                   for (int i = 0; i < dec.getNumPages(); i++) {
                        RenderedImage rm = new NullOpImage(dec.decodeAsRenderedImage(i), null, OpImage.OP_IO_BOUND, null);
                        thumImagesV.add(i, ImageHandler.createThumbnail(rm, 150));
              } catch (Exception e) {
                   e.printStackTrace();
    Main GUI Class
         private void buildGUI(File selectedFile) {
              configButtonPanel();
              configThumbnailPanel(selectedFile);//loading with out thumbnail image
              new TestThread(selectedFile).run();
         }

  • Lightroom problems with large folders of Raw images

    I have a problem processing folders with a large number of Raw images in Lightroom and the same folder is easily processed in Adobe Bridge (without much loading time). The folder locked up Lightroom more than once. The folder has Nikon D2X and D2H .NIF files and it has 850 images. I bought Lightroom to simplify the editing and sorting and sorting process but I often work out of large folders like this. Is Bridge a better answer for what I am looking for or am I doing something wrong (like not waiting long enough).

    It can take some ime for LR to import a large folder of RAWs. But once in it should be faster.
    How are you importing. Are they already on your HD, or coming from the camera with copy, or a cardreader?
    Don
    Don Ricklin, MacBook 1.83Ghz Duo 2 Core running 10.4.9 & Win XP, Pentax *ist D
    http://donricklin.blogspot.com/

  • Large number of movieclip slowdown the application

    Hi,
    I have a scrollbar function.In the content area i attached movieclips.Those moviclips have dynamic textfield.The scroll bar working for small list of contents.up to 500 or 600 moviclips
    When i attached 2000 or larger number of movieclip the Entire flash get slows. How to i fasten them with same number of movieclip.
    why flash player slow when attaching large number of movieclip? it takes near by 58% cpu usage at this time.What is the reason for this?
    Thanks,
    Siva

    imm also if you put much media inside the flash library (Audio,images, videos) those will be uploaded on Ram,,,
    so instead you can export your library as external library
    or
    use AS3 loader to load your media, immutablility will be fine here!! ^_^
    Goodluck

  • Best way to delete large number of records but not interfere with tlog backups on a schedule

    Ive inherited a system with multiple databases and there are db and tlog backups that run on schedules.  There is a list of tables that need a lot of records purged from them.  What would be a good approach to use for deleting the old records?
    Ive been digging through old posts, reading best practices etc, but still not sure the best way to attack it.
    Approach #1
    A one-time delete that did everything.  Delete all the old records, in batches of say 50,000 at a time.
    After each run through all the tables for that DB, execute a tlog backup.
    Approach #2
    Create a job that does a similar process as above, except dont loop.  Only do the batch once.  Have the job scheduled to start say on the half hour, assuming the tlog backups run every hour.
    Note:
    Some of these (well, most) are going to have relations on them.

    Hi shiftbit,
    According to your description, in my opinion, the type of this question is changed to discussion. It will be better and 
    more experts will focus on this issue and assist you. When delete large number of records from tables, you can use bulk deletions that it would not make the transaction log growing and runing out of disk space. You can
    take the table offline for maintenance, a complete reorganization is always best because it does the delete and places the table back into a pristine state. 
    For more information about deleting a large number of records without affecting the transaction log.
    http://www.virtualobjectives.com.au/sqlserver/deleting_records_from_a_large_table.htm
    Hope it can help.
    Regards,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

  • Unable to parse properly for a large number of requests

    Hi all
    in weblogic 9 .2 when a request comes which parses and reads a xml file then there is no problem. But at a time when a large number of request comes to read the xml file then it is behaving differently .One node it is unable to locate.The meaning of error is actually inappropriate
    java.lang.NullPointerException
         at com.sun.org.apache.xerces.internal.dom.ParentNode.nodeListItem(ParentNode.java:814)
         at com.sun.org.apache.xerces.internal.dom.ParentNode.item(ParentNode.java:828)
         at com.test.ObjectPersistanceXMLParser.getData(ObjectPersistanceXMLParser.java:46)
         at com.test.testservlet.doPost(testservlet.java:634)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:763)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:856)
         at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:225)
         at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:127)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:283)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:175)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3214)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:121)
         at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:1983)
         at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:1890)
         at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1344)
         at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
         at weblogic.work.ExecuteThread.run(ExecuteThread.java:181)
    But there is a node and is found while processing in a java program for the same xml file. But this occurs when a large number of request comes.
    Please suggest.

    Yes I parse xml as much as the request comes. and i do not want to synchronize here. The below code executes for each request. Do we have a solution here .
    DocumentBuilderFactory docBuilderFactory = DocumentBuilderFactory.newInstance();
    DocumentBuilder docBuilder = docBuilderFactory.newDocumentBuilder();
    Document configXML = docBuilder.parse(filePath);
    I have tried with DOMParser and follwowing method is applied
    setFeature( "http://apache.org/xml/features/dom/defer-node-expansion ",false).
    Please suggest.

Maybe you are looking for