Managed ODP: InvalidCastException when returning large number

I get an InvalidCastException "Specified cast is not valid" when trying to get a value bigger than Decimal.MaxValue as a scalar result from a select.
From the stack trace and inspection of the Oracle.ManagedDataAccess.dll code I would expect OracleDataReader to cast to OracleDecimal (or System.Numerics.BigInteger) instead of System.Decimal for large numbers.
Is this a bug?
Stack trace:
   at Oracle.ManagedDataAccess.Client.OracleDataReader.GetDecimal(Int32 i)
   at Oracle.ManagedDataAccess.Client.OracleDataReader.GetValue(Int32 i)
   at Oracle.ManagedDataAccess.Client.OracleCommand.ExecuteScalar()
   at Cora.Data.DBProvider.Tests.CoraParameterTest.BigIntegerOracleTest() in D:\svnroot\tech\lib\Cora.Data.DBProvider\Cora.Data.DBProvider.Tests\CoraParameterTest.cs:line 98
How to reproduce:
- create a table NUNITTEST_PARAM_SIMPLE with a column BIG of type NUMBER(38)
- run sample code
            using ( OracleConnection conn = new OracleConnection(ORACONN) )
                conn.Open();
                using ( OracleCommand cmd = conn.CreateCommand() )
                    cmd.CommandText = "insert into NUNITTEST_PARAM_SIMPLE (BIG) values (1234567890123456789012345678901234567)";
                    cmd.ExecuteNonQuery();
                using ( OracleCommand cmd = conn.CreateCommand() )
                    cmd.CommandText = "select BIG from NUNITTEST_PARAM_SIMPLE";
                    object result = cmd.ExecuteScalar();          // Exception
                conn.Close();

There is no safe type mapping for ExecuteScalar. It's an uncommon use case.
Even if there was, making the correct conversion implicitly would lead to more work for the application developer. Let's assume ODP.NET performs the "right" conversion. Now, the app needs to manipulate the data. What data type does it treat the returned value as: string or decimal? It could be either. That means twice as much code. One set deals with the data if a decimal was returned. The second deals with the data if a string is returned.
It would be much easier to write code for one scenario, rather than two. That is why ODP.NET safe type mapping is an explicit conversion the developer makes.

Similar Messages

  • Error "java.lang.OutOfMemoryError" When Returning Large Number of Docs

    In our SES implementation, we have a custom search interface that allows users to search for documents and then add them to a "shopping cart". Users add then to their shopping cart from search results, by adding docs one-by-one or an Add All option. Once they are done "shopping" they create a report.
    Here is the scenario...
    Users are saerching for documents and seeing that on page 1, there is 1 - 10 of about 300 results. They clicked Add All and want all 300 docs added to their cart.
    What we do under the covers is we execute another search and set the docs requested to 200. We get the array of docs, iterate over them and add the keys to a list. We found 200 docs at a time to be a safe number. However, there are still 100 docs that were not added to their cart and users want all 300 added. In other words, when they click Add All, they want to add all docs, 300, 500, 5000, etc.
    I set the "Maximum Number of Results" to 500 and found that I can safely add up to ~ 350 docs at one time. However, going past this point throws the following error:
    [SOAPException: faultCode=SOAP-ENV:Server; msg= [java.lang.OutOfMemoryError]]
         at oracle.search.query.webservice.client.OracleSearchService.makeSOAPCallRPC(OracleSearchService.java:941)
         at oracle.search.query.webservice.client.OracleSearchService.doOracleSearch(OracleSearchService.java:469)
    After this error is thrown, SES was unable to recover and searching would not work anymore, even return 10 docs at a time. We had to restart SES to resolve the issue.
    1. What is throwing this error? Is it the amount of XML being returned?
    2. What is the maximum number of results we can get back at any one time? Is it based on the amount of data being returned or the number of attributes?
    We are running 10.1.8 with plans to upgrade soon.
    Thanks in advance.

    I know it may be hard to throw away all this code, but consider using the jakarta fileupload component.
    I think it would simplify your code down to
    // Create a factory for disk-based file items
    FileItemFactory factory = new DiskFileItemFactory();
    // Create a new file upload handler
    ServletFileUpload upload = new ServletFileUpload(factory);
    // Parse the request
    List /* FileItem */ items = upload.parseRequest(request);
    // Process the uploaded items
    Iterator iter = items.iterator();
    while (iter.hasNext()) {
        FileItem item = (FileItem) iter.next();
        if (item.isFormField()) {
            processFormField(item);
        } else {
            // item is a file.  write it
            File saveFolder = application.getRealPath("/file");          
            File uploadedFile = new File(saveFolder, item.getName());
            item.write(uploadedFile);
    }Most of this code was hijacked from http://jakarta.apache.org/commons/fileupload/using.html
    Check it out. It will solve your memory problem by writing the file to disk temporarily if necessary.
    Cheers,
    evnafets

  • MBPro crashes when transfering large number of files

    i have a MBPro 2008 and recently when i transfer a large number of video files [over 5GB] my computer will suddendly crash in the middle of the transfer.
    i'm usually not using it while its transfering. This is a very recent problem.
    Sincerly,
    stef./.\

    How large is you HD and how much space do you have left?
    Try the following:
    Disconnect all devices from the computer then do the following:
    Boot up from your Snow Leopard Install DVD while holding down the "c" key.
    Select the language you wish to use, and then choose "Disk Utility" from the menu bar.
    Your HD should appear in a panel on the left hand side of the window which opens. Select it and then click on the "repair disk" button in the right hand section of the window.
    Once this process has been completed restart your computer and repair permissions directly from Disk Utility.
    If Disk Utility is unable to complete the repair, you will need to use a stronger 3rd party utility like DiskWarrior, Techtool PRO (not Deluxe) or Drive Genius.

  • Not receiving email when sending large number of records using a FM?

    Hi..
    I am using the function module " SO_DOCUMENT_SEND_API1 " to send email....
    When a single record is there.. or around 5-6 records are there... email is coming successfully...
    But when there are more records..say around 100, the email is not coming... I checked SOST transaction and the status there is in red..and the error message is " Internal error: SO_OBJECT_MIME_GET Exception: 2 ".........
    What could be the reason behind this problem....... ??
    I have another problem... my output has over 60 fields, but in the email which i am receiving has only around 10 fields... how to solve this problem...??
    Plz help...

    Well... right now i am tryin to get only the first 2 fields.. but even in this case... i am not getting the email if around 15 records are there......
    I am using the code which is given below which i found in SDN only......In this code.. data is getting selected from EKPO... i tried changing the number of rows getting selected.. and in this case, the attachement is coming as desired... but when i use the same code for my prog.. i am not getting the mail.. even if there are only 10 records or so...
    *& Report  ZT062108   ALV Header                                    *
    REPORT  zt062108.
    TABLES: ekko.
    PARAMETERS: p_email   TYPE somlreci1-receiver
                                      DEFAULT '<give email here>'.
    TYPES: BEGIN OF t_ekpo,
      ebeln TYPE ekpo-ebeln,
      ebelp TYPE ekpo-ebelp,
      aedat TYPE ekpo-aedat,
      matnr TYPE ekpo-matnr,
    END OF t_ekpo.
    DATA: it_ekpo TYPE STANDARD TABLE OF t_ekpo INITIAL SIZE 0,
          wa_ekpo TYPE t_ekpo.
    TYPES: BEGIN OF t_charekpo,
      ebeln(10) TYPE c,
      ebelp(5)  TYPE c,
      aedat(8)  TYPE c,
      matnr(18) TYPE c,
    END OF t_charekpo.
    DATA: wa_charekpo TYPE t_charekpo.
    DATA:   it_message TYPE STANDARD TABLE OF solisti1 INITIAL SIZE 0
                    WITH HEADER LINE.
    DATA:   it_attach TYPE STANDARD TABLE OF solisti1 INITIAL SIZE 0
                    WITH HEADER LINE.
    DATA:   t_packing_list LIKE sopcklsti1 OCCURS 0 WITH HEADER LINE,
            t_contents LIKE solisti1 OCCURS 0 WITH HEADER LINE,
            t_receivers LIKE somlreci1 OCCURS 0 WITH HEADER LINE,
            t_attachment LIKE solisti1 OCCURS 0 WITH HEADER LINE,
            t_object_header LIKE solisti1 OCCURS 0 WITH HEADER LINE,
            w_cnt TYPE i,
            w_sent_all(1) TYPE c,
            w_doc_data LIKE sodocchgi1,
            gd_error    TYPE sy-subrc,
            gd_reciever TYPE sy-subrc.
    *START_OF_SELECTION
    START-OF-SELECTION.
    *   Retrieve sample data from table ekpo
      PERFORM data_retrieval.
    *   Populate table with detaisl to be entered into .xls file
      PERFORM build_xls_data_table.
    *END-OF-SELECTION
    END-OF-SELECTION.
    * Populate message body text
      perform populate_email_message_body.
    * Send file by email as .xls speadsheet
      PERFORM send_file_as_email_attachment
                                   tables it_message
                                          it_attach
                                    using p_email
                                          'Example .xls documnet attachment'
                                          'TXT'
                                          'filename'
                                 changing gd_error
                                          gd_reciever.
    *   Instructs mail send program for SAPCONNECT to send email(rsconn01)
      PERFORM initiate_mail_execute_program.
    *&      Form  DATA_RETRIEVAL
    *       Retrieve data form EKPO table and populate itab it_ekko
    FORM data_retrieval.
      SELECT ebeln ebelp aedat matnr
       UP TO 1000 ROWS
        FROM ekpo
        INTO TABLE it_ekpo.
    ENDFORM.                    " DATA_RETRIEVAL
    *&      Form  BUILD_XLS_DATA_TABLE
    *       Build data table for .xls document
    FORM build_xls_data_table.
    *  CONSTANTS: con_cret TYPE x VALUE '0D'.  "OK for non Unicode
    *             con_tab TYPE x VALUE '09'.   "OK for non Unicode
    *If you have Unicode check active in program attributes thnen you will
    *need to declare constants as follows
    *class cl_abap_char_utilities definition load.
    constants:
        con_tab  type c value cl_abap_char_utilities=>HORIZONTAL_TAB,
        con_cret type c value cl_abap_char_utilities=>CR_LF.
      CONCATENATE 'EBELN' 'EBELP' 'AEDAT' 'MATNR'
             INTO it_attach SEPARATED BY con_tab.
      CONCATENATE con_cret it_attach  INTO it_attach.
      APPEND  it_attach.
      LOOP AT it_ekpo INTO wa_charekpo.
        CONCATENATE wa_charekpo-ebeln wa_charekpo-ebelp
                    wa_charekpo-aedat wa_charekpo-matnr
               INTO it_attach SEPARATED BY con_tab.
        CONCATENATE con_cret it_attach  INTO it_attach.
        APPEND  it_attach.
      ENDLOOP.
    ENDFORM.                    " BUILD_XLS_DATA_TABLE
    *&      Form  SEND_FILE_AS_EMAIL_ATTACHMENT
    *       Send email
    FORM send_file_as_email_attachment tables pit_message
                                              pit_attach
                                        using p_email
                                              p_mtitle
                                              p_format
                                              p_filename
                                              p_attdescription
                                              p_sender_address
                                              p_sender_addres_type
                                     changing p_error
                                              p_reciever.
      DATA: ld_error    TYPE sy-subrc,
            ld_reciever TYPE sy-subrc,
            ld_mtitle LIKE sodocchgi1-obj_descr,
            ld_email LIKE  somlreci1-receiver,
            ld_format TYPE  so_obj_tp ,
            ld_attdescription TYPE  so_obj_nam ,
            ld_attfilename TYPE  so_obj_des ,
            ld_sender_address LIKE  soextreci1-receiver,
            ld_sender_address_type LIKE  soextreci1-adr_typ,
            ld_receiver LIKE  sy-subrc.
      ld_email   = p_email.
      ld_mtitle = p_mtitle.
      ld_format              = p_format.
      ld_attdescription      = p_attdescription.
      ld_attfilename         = p_filename.
      ld_sender_address      = p_sender_address.
      ld_sender_address_type = p_sender_addres_type.
    * Fill the document data.
      w_doc_data-doc_size = 1.
    * Populate the subject/generic message attributes
      w_doc_data-obj_langu = sy-langu.
      w_doc_data-obj_name  = 'SAPRPT'.
      w_doc_data-obj_descr = ld_mtitle .
      w_doc_data-sensitivty = 'F'.
    * Fill the document data and get size of attachment
      CLEAR w_doc_data.
      READ TABLE it_attach INDEX w_cnt.
      w_doc_data-doc_size =
         ( w_cnt - 1 ) * 255 + STRLEN( it_attach ).
      w_doc_data-obj_langu  = sy-langu.
      w_doc_data-obj_name   = 'SAPRPT'.
      w_doc_data-obj_descr  = ld_mtitle.
      w_doc_data-sensitivty = 'F'.
      CLEAR t_attachment.
      REFRESH t_attachment.
      t_attachment[] = pit_attach[].
    * Describe the body of the message
      CLEAR t_packing_list.
      REFRESH t_packing_list.
      t_packing_list-transf_bin = space.
      t_packing_list-head_start = 1.
      t_packing_list-head_num = 0.
      t_packing_list-body_start = 1.
      DESCRIBE TABLE it_message LINES t_packing_list-body_num.
      t_packing_list-doc_type = 'RAW'.
      APPEND t_packing_list.
    * Create attachment notification
      t_packing_list-transf_bin = 'X'.
      t_packing_list-head_start = 1.
      t_packing_list-head_num   = 1.
      t_packing_list-body_start = 1.
      DESCRIBE TABLE t_attachment LINES t_packing_list-body_num.
      t_packing_list-doc_type   =  ld_format.
      t_packing_list-obj_descr  =  ld_attdescription.
      t_packing_list-obj_name   =  ld_attfilename.
      t_packing_list-doc_size   =  t_packing_list-body_num * 255.
      APPEND t_packing_list.
    * Add the recipients email address
      CLEAR t_receivers.
      REFRESH t_receivers.
      t_receivers-receiver = ld_email.
      t_receivers-rec_type = 'U'.
      t_receivers-com_type = 'INT'.
      t_receivers-notif_del = 'X'.
      t_receivers-notif_ndel = 'X'.
      APPEND t_receivers.
      CALL FUNCTION 'SO_DOCUMENT_SEND_API1'
           EXPORTING
                document_data              = w_doc_data
                put_in_outbox              = 'X'
                sender_address             = ld_sender_address
                sender_address_type        = ld_sender_address_type
                commit_work                = 'X'
           IMPORTING
                sent_to_all                = w_sent_all
           TABLES
                packing_list               = t_packing_list
                contents_bin               = t_attachment
                contents_txt               = it_message
                receivers                  = t_receivers
           EXCEPTIONS
                too_many_receivers         = 1
                document_not_sent          = 2
                document_type_not_exist    = 3
                operation_no_authorization = 4
                parameter_error            = 5
                x_error                    = 6
                enqueue_error              = 7
                OTHERS                     = 8.
    * Populate zerror return code
      ld_error = sy-subrc.
    * Populate zreceiver return code
      LOOP AT t_receivers.
        ld_receiver = t_receivers-retrn_code.
      ENDLOOP.
    ENDFORM.
    *&      Form  INITIATE_MAIL_EXECUTE_PROGRAM
    *       Instructs mail send program for SAPCONNECT to send email.
    FORM initiate_mail_execute_program.
      WAIT UP TO 2 SECONDS.
      SUBMIT rsconn01 WITH mode = 'INT'
                    WITH output = 'X'
                    AND RETURN.
    ENDFORM.                    " INITIATE_MAIL_EXECUTE_PROGRAM
    *&      Form  POPULATE_EMAIL_MESSAGE_BODY
    *        Populate message body text
    form populate_email_message_body.
      REFRESH it_message.
      it_message = 'Please find attached a list test ekpo records'.
      APPEND it_message.
    endform.                    " POPULATE_EMAIL_MESSAGE_BODY

  • CLR Procedure and returning large number of rows

    I have a CLR stored procedure coded in C# that retrieves data from a web service, and returns that data using SendResultsStart/SendResultsRow/SendResultsEnd. This all works fine, except when the data from the web service is tens or even thousands of records.
    The code itself takes about 3 minutes on average to do all it's work with around 50000-60000 records, but the procedure does not return in SSMS for about another 10-15 minutes, during which time the CPU and memory usage go up significantly.
    To rule out any of the CLR code as the culprit, I created a very simple CLR procedure that just loops to return 100000 records with int and nvarchar(256) fields with the current count, and "ABC" followed by the count.  Here is the code:
    [Microsoft.SqlServer.Server.SqlProcedure]
    public static void ABC()
    System.Diagnostics.Stopwatch ExecuteTimer = System.Diagnostics.Stopwatch.StartNew();
    SqlMetaData[] ResultMetaData = new SqlMetaData[2];
    ResultMetaData[0] = new SqlMetaData("count", SqlDbType.Int);
    ResultMetaData[1] = new SqlMetaData("text", SqlDbType.NVarChar, 256);
    SqlContext.Pipe.SendResultsStart(new SqlDataRecord(ResultMetaData));
    for (int x = 0; x < 100000; x++)
    SqlDataRecord ResultItem = new SqlDataRecord(ResultMetaData);
    ResultItem.SetValue(0, x);
    ResultItem.SetValue(1, "ABC" + x.ToString());
    SqlContext.Pipe.SendResultsRow(ResultItem);
    SqlContext.Pipe.SendResultsEnd();
    TimeSpan ExecTime = ExecuteTimer.Elapsed;
    SqlContext.Pipe.Send("Elapsed Time: " + ExecTime.Minutes.ToString() + ":" + ExecTime.Seconds.ToString() + "." + ExecTime.Milliseconds.ToString());
    I then executed procedure ABC in SSMS, and it took 21 minutes to return.  All of the data rows were visible after just a couple of seconds, but the query continued to run as the CPU and memory went up again.
    Is this really how long it should take to return 100000 rows, or am I missing something?  Is there a better approach than using SendResultsStart/SendResultsRow/SendResultsEnd?
    I've googled this to death and haven't found anything that helped or even explained why this is.
    I would greatly appreciate any suggestions or alternate methods to achieve this faster.
    Thanks!
    Alex

    When you create a new object, space on the garbage-collected heap is allocated for that object, and the address will be stored in a reference. Some time later, there will no longer be any references that hold the address of the allocated object. It doesn't
    matter whether the reference count went to 0 because the reference was set to null or because the reference was on the stack and is no longer in lexical scope, the end result is the same: the garbage collector will, at some point in time, have to perform the
    book-keeping operations necessary to identify that the space allocated for that now-unreferenced object can be re-used. When, on the other hand, you only create a single SqlDataRecord object and hold onto the reference, all of the book-keeping operations
    associated with creating 100,000 objects are eliminated. This is why the documentation for the SqlDataReader class advises that:
    When writing common language runtime (CLR) applications, you should re-use existing
    SqlDataRecord objects instead of creating new ones every time. Creating many new
    SqlDataRecord objects could severely deplete memory and adversely affect performance.

  • When loading large number of images, images begin to fail to load in IE11 with DOM: 7009 error (unable to decode) in console

    When loading many images on a single page in IE (reproduced in IE11), some of them begin to fail to load, and have something similar to the following warning in the console:
    DOM7009: Unable to decode image at URL: '[some unique url]'.
    When I look at the network traffic, there does appear to valid responses received for each of these images from the server. It's not always the same images each time. If I use the dev tools to force the image to reload (example: I update the url to include some some extraneous url parameter "&test=1"), it loads/renders normally without error. I've reproduced this behavior with different types of images (jpegs/pngs; example png included below). Apologies for the cross-posting. I wasn't sure what the particular etiquette was for that - I've also posted this issue on StackOverflow and MS Connect under the same title. If one of those is a more appropriate place for this matter, please let me know and I'll close this issue.

    yup, i have encountered this problem as well. I'm loading an array of image frames for a movie, and i'm using preloadJS (javascript preloading library) and not displaying the images upfront. 

  • ActiveX Printer Error when print large number of page

    Hi,
    I have this problem when developing the report on my project.
    I developed the Crystal Report embedded in Visual Studio 2008 to be displayed in web form.
    I'm using .NET and the operating system of the development computer is Windows 7 32 bit.
    The problem is, I have a quite complex report displaying the statement for the customer which is result in 209 pages.
    When the report has successfully loaded, I simply click the print icon and it comes up with the error "A communication error occurred. Printing will be stopped"
    But, after that...I tried something different, when the report has successfully loaded, I went through the report by moving for every 30 pages until it reaches the last page and click the print icon. and it prints without any errors!
    Anybody knows what caused this?
    By the way, I'm using the ActiveX printing mode for the Crystal Report.
    Thanks
    -Marry-

    I'm thinking there is something not processed in the report so forcing the report to process everything up to the last page before printing may help. Try the following code before sending the report to the viewer:
    Dim prc As New CrystalDecisions.Shared.ReportPageRequestContext
    rptDoc.FormatEngine.GetLastPageNumber(prc)
    This may slow down the time it takes for the report to come up as the whole report will need to be processed, formatted, etc.
    Also, make sure you have SP 1 for CR 10.5:
    https://smpdl.sap-ag.de/~sapidp/012002523100009351512008E/crbasic2008sp1.exe
    Ludek
    Follow us on Twitter http://twitter.com/SAPCRNetSup
    Got Enhancement ideas? Try the [SAP Idea Place|https://ideas.sap.com/community/products_and_solutions/crystalreports]

  • I-bot not emailing when report returns large result set..

    Hi,
    I am trying to set up an i-bot to run daily and email the results to the user. Assuming the report in question is Report_A.
    Report_A returns around 60000 rows of data without any filter condition. When I tried to set up thei-bot for Report_A (No filter conditions on the report) the ibot is publishing results to dashboard but is not delivering via email. When I introduce a filter in Report_A to reduce the data returned then everything works fine and email is being sent out successfully.
    So
    1) Is there a size limit for i-bots to deliver by email?
    2) Is there a way to increase the limits if any so the report can be emailed even when returning large result sets?
    Please let me know.

    Sorry for late reply
    Below is the log file for one of the i-bots. Now I am getting an error message "***kmsgPortalGoRequestHasBeenCancelled: message text not found ***" and the i-bot alert message shows as "Cancelled".
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:04.551
    [nQSError: 77006] Oracle BI Presentation Server Error: A fatal error occurred while processing the request. The server responded with: ***kmsgPortalGoRequestHasBeenCancelled: message text not found ***
    Error Codes: YLKKAV7S
    Error Codes: AGEGTYVF
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:04.553
    iBotID: /shared/_ibots/common/TM/Claims Report
    ...Trying iBot Get Response Content loop again.
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:04.554
    ... Sleeping for 8 seconds.
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:12.642
    [nQSError: 77006] Oracle BI Presentation Server Error: A fatal error occurred while processing the request. The server responded with: ***kmsgPortalGoRequestHasBeenCancelled: message text not found ***
    Error Codes: YLKKAV7S
    Error Codes: AGEGTYVF
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:12.644
    iBotID: /shared/_ibots/common/TM/Claims Report
    ...Trying iBot Get Response Content loop again.
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:12.644
    ... Sleeping for 6 seconds.
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:18.730
    [nQSError: 77006] Oracle BI Presentation Server Error: A fatal error occurred while processing the request. The server responded with: ***kmsgPortalGoRequestHasBeenCancelled: message text not found ***
    Error Codes: YLKKAV7S
    Error Codes: AGEGTYVF
    +++ ThreadID: f3c6cb90 : 2010-12-17 23:55:18.734
    iBotID: /shared/_ibots/common/TM/Claims Report
    Exceeded number of request retries.

  • Best way to delete large number of records but not interfere with tlog backups on a schedule

    Ive inherited a system with multiple databases and there are db and tlog backups that run on schedules.  There is a list of tables that need a lot of records purged from them.  What would be a good approach to use for deleting the old records?
    Ive been digging through old posts, reading best practices etc, but still not sure the best way to attack it.
    Approach #1
    A one-time delete that did everything.  Delete all the old records, in batches of say 50,000 at a time.
    After each run through all the tables for that DB, execute a tlog backup.
    Approach #2
    Create a job that does a similar process as above, except dont loop.  Only do the batch once.  Have the job scheduled to start say on the half hour, assuming the tlog backups run every hour.
    Note:
    Some of these (well, most) are going to have relations on them.

    Hi shiftbit,
    According to your description, in my opinion, the type of this question is changed to discussion. It will be better and 
    more experts will focus on this issue and assist you. When delete large number of records from tables, you can use bulk deletions that it would not make the transaction log growing and runing out of disk space. You can
    take the table offline for maintenance, a complete reorganization is always best because it does the delete and places the table back into a pristine state. 
    For more information about deleting a large number of records without affecting the transaction log.
    http://www.virtualobjectives.com.au/sqlserver/deleting_records_from_a_large_table.htm
    Hope it can help.
    Regards,
    Sofiya Li
    Sofiya Li
    TechNet Community Support

  • Large number of movieclip slowdown the application

    Hi,
    I have a scrollbar function.In the content area i attached movieclips.Those moviclips have dynamic textfield.The scroll bar working for small list of contents.up to 500 or 600 moviclips
    When i attached 2000 or larger number of movieclip the Entire flash get slows. How to i fasten them with same number of movieclip.
    why flash player slow when attaching large number of movieclip? it takes near by 58% cpu usage at this time.What is the reason for this?
    Thanks,
    Siva

    imm also if you put much media inside the flash library (Audio,images, videos) those will be uploaded on Ram,,,
    so instead you can export your library as external library
    or
    use AS3 loader to load your media, immutablility will be fine here!! ^_^
    Goodluck

  • Unable to parse properly for a large number of requests

    Hi all
    in weblogic 9 .2 when a request comes which parses and reads a xml file then there is no problem. But at a time when a large number of request comes to read the xml file then it is behaving differently .One node it is unable to locate.The meaning of error is actually inappropriate
    java.lang.NullPointerException
         at com.sun.org.apache.xerces.internal.dom.ParentNode.nodeListItem(ParentNode.java:814)
         at com.sun.org.apache.xerces.internal.dom.ParentNode.item(ParentNode.java:828)
         at com.test.ObjectPersistanceXMLParser.getData(ObjectPersistanceXMLParser.java:46)
         at com.test.testservlet.doPost(testservlet.java:634)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:763)
         at javax.servlet.http.HttpServlet.service(HttpServlet.java:856)
         at weblogic.servlet.internal.StubSecurityHelper$ServletServiceAction.run(StubSecurityHelper.java:225)
         at weblogic.servlet.internal.StubSecurityHelper.invokeServlet(StubSecurityHelper.java:127)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:283)
         at weblogic.servlet.internal.ServletStubImpl.execute(ServletStubImpl.java:175)
         at weblogic.servlet.internal.WebAppServletContext$ServletInvocationAction.run(WebAppServletContext.java:3214)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:321)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:121)
         at weblogic.servlet.internal.WebAppServletContext.securedExecute(WebAppServletContext.java:1983)
         at weblogic.servlet.internal.WebAppServletContext.execute(WebAppServletContext.java:1890)
         at weblogic.servlet.internal.ServletRequestImpl.run(ServletRequestImpl.java:1344)
         at weblogic.work.ExecuteThread.execute(ExecuteThread.java:209)
         at weblogic.work.ExecuteThread.run(ExecuteThread.java:181)
    But there is a node and is found while processing in a java program for the same xml file. But this occurs when a large number of request comes.
    Please suggest.

    Yes I parse xml as much as the request comes. and i do not want to synchronize here. The below code executes for each request. Do we have a solution here .
    DocumentBuilderFactory docBuilderFactory = DocumentBuilderFactory.newInstance();
    DocumentBuilder docBuilder = docBuilderFactory.newDocumentBuilder();
    Document configXML = docBuilder.parse(filePath);
    I have tried with DOMParser and follwowing method is applied
    setFeature( "http://apache.org/xml/features/dom/defer-node-expansion ",false).
    Please suggest.

  • I have a large number of PDFs documents stored on a windows pcthat I would like to access on my iPad mini even when not cBooking referenceonnected to a network.

    i have a large number of PDFs documents stored on a windows pc that I would like to access on my iPad mini even when not connected  to a network. Does anyone know a solution. Eg maybe card reader and a app.

    Those devices have mixed results depending on the application. If the application does not support the device, then you are out of luck. There is no file manager in IOS to allow you to move files between folders because there are no folders that you can access. An app has to specifically support a storage location to access files.
    It's not that bad. The app has to have the ability to accept data from another app.  The latest memory dongles come with an ios app to access the external storage.  You then transfer the pdf from dongle's app to your app.  If your app can accept data from dropbox, it will be able to accept data from the memory dongle app.
    Other solutions:
    == "GoodReaderUSB is a practical and useful application whose main purpose is to help users transfer files and folders from their mobile device to PC effortlessly. It helps them to move and backup important data from their Apple device via a USB cable."
    http://www.softpedia.com/get/IPOD-TOOLS/Multimedia-IPOD-tools/GoodReaderUSB.shtm l
    == "iExplorer's disk mounting features allow you to use your iPhone, iPod or iPad like a USB flash drive."
    http://www.macroplant.com/iexplorer/

  • Sender File Adapter does not get files when there is a large number ( 300)

    We are on XI 7.0 (SAPKB70015)
    We have an issue picking up a large number of files from a particular server. We are trying to pickup the files based on a wildcard e.g. test
    - we can retreive up to 200 files from the same server (using the same config) no problem
    - we can pull lots of files (>3000) files from other servers (using the same config) no problem
    - we can write an mget command (mget test) to reteive the files no problem
    So the only issue is retrieving a large number of files from a parfticular server. We see no errors in the RWB.
    To get this sorted out we need to find the exact commmands that XI executes to attempt to retrieve the files. Can anyone point me in the right direction on how to find this.
    Thanks
    Richard

    It turned out that this was caused by a limitation on the ftp server.
    The file adapter uses the dir command to get a listing of the files matching a specific mask. The dir command does not return any files if there are a large number (the ls command works fine but the file adapter uses dir instead of ls).
    sap told me that 'According to this FTP server is no compliant with RFC 959 and that is
    why DIR command returns an error.'
    We were able to get this resolved by applying a patch on the ftp server.
    Thanks
    Richard

  • Generalizing the return page number when linking to a page in another app

    Hello.
    I know how to link to a page in another app within the same workspace. However, my problem arises when the user tries to go back to the page in the first app.
    Is there a way to "generalize" the "return path". I do not want to hard-code the return page number within a "back link" element on the target page. I was wondering if there is some sort of substitution string I can use in the return http address that automatically "knows" what the returning page number is. Hope I'm making sense here.
    Thanks for any help/advice.

    Funky:
    I have tried some work but no success
    I don't know what you are asking of me. I have given detailed step-by-step instructions on how to implement what I suggested earlier on in this thread. Look at my postings on 9/13/2006. What part are you having trouble with?
    Could you be more specific?
    I'm afraid I am going to have to ask you the same question.
    if page is not submited (called through Tab)
    Apex standard tabs do submit the page and fire after-submit computations defined on the calling page before branching to the "current" page for the tab you click on.

  • Problem with function, need to return a very large number

    I have a function that is used to derive a number by using a defined algorithm. We have tested it many times in our 'development' region and it works. We moved it to our 'UAT' region and now the function is giving an error. The data is the same in both oracle instances. So.... dba's are telling me it is the function. At this point, I don't know what to do.
    Can I declare "RESULT" in a function so that it can return a number this big: 2.00023880306E24?
    Oracle returns -6502 error in the deriveNewNum function.
    Here is the error:
    *** ERROR in deriveNewNum.fnc
    SQLCODE: -6502
    ORA-06502: PL/SQL: numeric or value error: number precision too large
    BEGIN TF1ID; END;
    ERROR at line 1:
    ORA-06503: PL/SQL: Function returned without value
    ORA-06512: at DERIVENEWNUM", line 93
    ORA-06512: at line 1

    Hi,
    Did you look up that error in the error messages manual?
    http://download.oracle.com/docs/cd/B28359_01/server.111/b28278/e4100.htm#sthref1889
    Is there something there that you didn't understand?
    <h3>ORA-06502: PL/SQL: numeric or value errorstring </h3>
    Cause: An arithmetic, numeric, string, conversion, or constraint error occurred. For example, this error occurs if an attempt is made to assign the value NULL to a variable declared NOT NULL, or if an attempt is made to assign an integer larger than 99 to a variable declared NUMBER(2).
    Action: Change the data, how it is manipulated, or how it is declared so that values do not violate constraints.
    If you declare variables like this:
    x     NUMBER;
    y     NUMBER (20, 10);there woul be no problem assigning a value like 2E24 to x, but there would be if you tried it with y.
    How are your varaibles declared.
    Whenever you have a problem with your code, post your code.
    Always say what version of Oracle you're using.

Maybe you are looking for

  • Wont open to windows..diagnostic gives many error codes

    Won't open to windows, did diagnostic   hard disc error codes follow....u4wc59-62w7as-mfkt01-60qlo3  hard disc failure.  Does this mean hard drive is shot?  Or needs new software program?    Thanks!

  • Status display in monitor

    Hi experts, In the consolidation monitor, I've got consolidation group hierarchy (company) called H1, and bellow all companies. When I execute the function of data collection at the consolidation group level (H1) in the update mode, it executes the f

  • Faulting Application JFMerge.exe

    Hi all, Our servers have just started generating an application error the info looks like this 'Reporting queued error: Faulting application JFMerge.exe, version 5.5.0.41.42, faulting module JFMerge.exe, version 5.5.5.0.41.42, fault address ...' Anyo

  • Why there are big differences in the prices of spare parts (hinge) ..?? particular egypt ibm

    I have Lenovo 3000 N200 0769 with problem in left hinge lcd , the number of it i think (41W5214) 15.4 lcd. lenove egypt (IBM) was told to me cost 100 $ but in Many sites (hinge) cost betwen 19 $ , 30$ , 80 $ i want to understand why there are a big d

  • Inability of FireFox to verify security certificates.

    The warning below shows up consistently on every attempt to access the web, whether directly or by link such as in an e-mail: This includes sites regularly accessed including Google and Mozilla. This Connection is Untrusted You have asked Firefox to