Error:  Debug Component exceeds maximum size (65535 bytes)

Hi All,
It would seem that I have come across a limitation in the CAP file format for Java Card 2.2.1. When I run the Sun 2.2.1 converter I get the following output:
Converting oncard package
Java Card 2.2.1 Class File Converter, Version 1.3
Copyright 2003 Sun Microsystems, Inc. All rights reserved. Use is subject to license terms.
error:  Debug Component exceeds maximum size (65535 bytes).
Cap file generation failed.
conversion completed with 1 errors and 0 warnings.This also happens with JCDK 2.2.2. Is there any way around this limitation? I can instruct the compiler to not generate some of the debug output such as variable attributes, but this makes the debugger less than useful. I could not find anything in the JCVM spec for JC2.2.1 that mentions this limitation (I could have missed it very easily if it is there). Is this a limitation that will be lifted in new versions of the standard? This is tarting to be a bit of a problem as our code base has outgrown our development tools (no more debugger). Compiling and running without debug information works fine.
Any information on this would be much appreciated.
Cheers,
Shane

For those that are interested, I found this in the Java Card VM spec. u1 and u2 are 1 and 2 byte unsigned values respectively. It would appear that it is a limitation of the CAP file format. We have managed to work around this problem (parsing the debug component of the cap file was very informative) by shortening package names, reducing exception handling, and removing classes that we could do without.
It looks like this is removed for JC3.0 Connected Edition does not have this issue as it does not have CAP files, but JC2.2.2 and JC3.0 Classic Edition have this same issue. It would be nice if there was a converter/debugger combination that removed this limitation. If anyone knows of such a combination, I am all ears! I know for a fact that JCOP does not :(
Cheers,
Shane
*6.14 Debug Component*
This section specifies the format for the Debug Component. The Debug Component contains all the metadata necessary for debugging a package on a suitably instrumented Java Card virtual machine. It is not required for executing Java Card programs in a non-debug environment.
The Debug Component references the Class Component (Section 6.8 "Class Component”), Method Component (Section 6.9 "Method Component”), and Static Field Component (Section 6.10 "Static Field Component”). No components reference the Debug Component.
The Debug Component is represented by the following structure:
{code}
debug_component {
u1 tag
u2 size
u2 string_count
utf8_info strings_table[string_count]
u2 package_name_index
u2 class_count
class_debug_info classes[class_count]
{code}
The items in the debug_component structure are defined as follows:
*tag* tag item has the value COMPONENT_Debug (12).
*size* The number of bytes in the component, excluding the tag and size items. The value of size must be greater than zero.

Similar Messages

  • If Dimension exceeds maximum size then what will we do???

    Hi Experts,
                   If Dimension exceeds maximum size then what will we do???
                   My dought was how to increase the dimension size in SSAS 2008 R2???
           i am using SQL SERVER 2008 R2.
                   i had faced this question in one of the interview.so,Could you explain with an example.
    Best Regards,
    sirikumar

    You can't exceed the maximum, else you get error. The maximum is a huge number:
    Object
    Maximum sizes/numbers
    Databases in an instance
    2^31-1 = 2,147,483,647
    Dimensions in a database
    2^31-1 = 2,147,483,647
    Attributes in a dimension
    2^31-1 = 2,147,483,647
    Members in a dimension attribute
    2^31-1 = 2,147,483,647
    User-defined hierarchies in a dimension
    2^31-1 = 2,147,483,647
    Levels in a user-defined hierarchy
    2^31-1 = 2,147,483,647
    Cubes in a database
    2^31-1 = 2,147,483,647
    LINK:
    Maximum Capacity Specifications (Analysis Services)
    Kalman Toth Database & OLAP Architect
    SELECT Video Tutorials 4 Hours
    New Book / Kindle: Exam 70-461 Bootcamp: Querying Microsoft SQL Server 2012

  • Errors with Queue exceed maximum capacity of: '65536' elements

    I faced this issue recently , where the messaging bridges from an external domain was not able to connect my my domain.
    When i checked the logs i found one of the server with the below error.
    ####<Feb 28, 2011 7:31:23 PM GMT> <Warning> <RMI> <dyh75a03> <managed06_orneop02> <ExecuteThread: '2' for queue: 'weblogic.socket.Muxer'> <<WLS Kernel>> <> <
    BEA-080003> <RuntimeException thrown by rmi server: weblogic.rmi.internal.BasicServerRef@a - hostID: '4275655823338432757S:dywlms06-orneop02.neo.openreach.co
    .uk:[61007,61007,-1,-1,61007,-1,-1,0,0]:dywlms01-orneop02.neo.openreach.co.uk:61002,dywlms02-orneop02.neo.openreach.co.uk:61003,dywlms03-orneop02.neo.openrea
    ch.co.uk:61004,dywlms04-orneop02.neo.openreach.co.uk:61005,dywlms05-orneop02.neo.openreach.co.uk:61006,dywlms06-orneop02.neo.openreach.co.uk:61007,dywlms07-o
    rneop02.neo.openreach.co.uk:61008,dywlms08-orneop02.neo.openreach.co.uk:61009,dywlms09-orneop02.neo.openreach.co.uk:61010:orneop02:managed06_orneop02', oid:
    '10', implementation: 'weblogic.transaction.internal.CoordinatorImpl@f6ede1'
    weblogic.utils.UnsyncCircularQueue$FullQueueException: Queue exceed maximum capacity of: '65536' elements.
    weblogic.utils.UnsyncCircularQueue$FullQueueException: Queue exceed maximum capacity of: '65536' elements
    at weblogic.utils.UnsyncCircularQueue.expandQueue(UnsyncCircularQueue.java:72)
    at weblogic.utils.UnsyncCircularQueue.put(UnsyncCircularQueue.java:94)
    at weblogic.kernel.ExecuteThreadManager.execute(ExecuteThreadManager.java:374)
    at weblogic.kernel.Kernel.execute(Kernel.java:345)
    at weblogic.rmi.internal.BasicServerRef.dispatch(BasicServerRef.java:312)
    at weblogic.rjvm.RJVMImpl.dispatchRequest(RJVMImpl.java:1113)
    at weblogic.rjvm.RJVMImpl.dispatch(RJVMImpl.java:1031)
    at weblogic.rjvm.ConnectionManagerServer.handleRJVM(ConnectionManagerServer.java:225)
    at weblogic.rjvm.ConnectionManager.dispatch(ConnectionManager.java:805)
    at weblogic.rjvm.t3.T3JVMConnection.dispatch(T3JVMConnection.java:782)
    at weblogic.socket.SocketMuxer.readReadySocketOnce(SocketMuxer.java:705)
    at weblogic.socket.SocketMuxer.readReadySocket(SocketMuxer.java:651)
    at weblogic.socket.PosixSocketMuxer.processSockets(PosixSocketMuxer.java:123)
    at weblogic.socket.SocketReaderRequest.execute(SocketReaderRequest.java:32)
    at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:224)
    at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:183)
    >
    As a workaround , i had restart this server. After this bridges from ext domain was working.
    Has any one come across this issue before.
    Regards
    Deepak

    Found this issue again in one of the domains....the threads dumps is showing deadlocks.
    Found one Java-level deadlock:
    =============================
    "ExecuteThread: '3' for queue: 'weblogic.kernel.Default'":
    waiting to lock monitor 0868dea4 (object a4e99be8, a weblogic.transaction.internal.TransactionLoggerImpl$IOBuffer),
    which is held by "ExecuteThread: '6' for queue: 'weblogic.kernel.Default'"
    "ExecuteThread: '6' for queue: 'weblogic.kernel.Default'":
    waiting to lock monitor 00255304 (object b23be050, a weblogic.transaction.internal.ServerTransactionImpl),
    which is held by "ExecuteThread: '0' for queue: 'weblogic.kernel.Non-Blocking'"
    "ExecuteThread: '0' for queue: 'weblogic.kernel.Non-Blocking'":
    waiting to lock monitor 086677a4 (object a4e99c00, a weblogic.transaction.internal.TransactionLoggerImpl$IOBuffer),
    which is held by "ExecuteThread: '3' for queue: 'weblogic.kernel.Default'"
    Java stack information for the threads listed above:
    ===================================================
    "ExecuteThread: '3' for queue: 'weblogic.kernel.Default'":
         at weblogic.transaction.internal.TransactionLoggerImpl$LogDisk.release(TransactionLoggerImpl.java:1323)
         - waiting to lock <a4e99be8> (a weblogic.transaction.internal.TransactionLoggerImpl$IOBuffer)
         - locked <a4e99c00> (a weblogic.transaction.internal.TransactionLoggerImpl$IOBuffer)
         at weblogic.transaction.internal.TransactionLoggerImpl.release(TransactionLoggerImpl.java:389)
         at weblogic.transaction.internal.ServerTransactionImpl.releaseLog(ServerTransactionImpl.java:2767)
         at weblogic.transaction.internal.ServerTransactionManagerImpl.remove(ServerTransactionManagerImpl.java:1466)
         at weblogic.transaction.internal.ServerTransactionImpl.afterCommittedStateHousekeeping(ServerTransactionImpl.java:2645)
         at weblogic.transaction.internal.ServerTransactionImpl.setCommitted(ServerTransactionImpl.java:2669)
         at weblogic.transaction.internal.ServerTransactionImpl.localCommit(ServerTransactionImpl.java:1875)
         at weblogic.transaction.internal.ServerTransactionImpl.localCommit(ServerTransactionImpl.java:1163)
         at weblogic.transaction.internal.SubCoordinatorImpl.startCommit(SubCoordinatorImpl.java:274)
         at weblogic.transaction.internal.CoordinatorImpl_WLSkel.invoke(Unknown Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:492)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:435)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:147)
         at weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:430)
         at weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:35)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:224)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:183)
    "ExecuteThread: '6' for queue: 'weblogic.kernel.Default'":
         at weblogic.transaction.internal.ServerTransactionImpl.onDisk(ServerTransactionImpl.java:836)
         - waiting to lock <b23be050> (a weblogic.transaction.internal.ServerTransactionImpl)
         at weblogic.transaction.internal.TransactionLoggerImpl$LogDisk.write(TransactionLoggerImpl.java:1252)
         - locked <a4e99be8> (a weblogic.transaction.internal.TransactionLoggerImpl$IOBuffer)
         - locked <a4e99bb8> (a weblogic.transaction.internal.TransactionLoggerImpl$LogDisk)
         at weblogic.transaction.internal.TransactionLoggerImpl.flushLog(TransactionLoggerImpl.java:614)
         at weblogic.transaction.internal.TransactionLoggerImpl.store(TransactionLoggerImpl.java:305)
         at weblogic.transaction.internal.ServerTransactionImpl.log(ServerTransactionImpl.java:1850)
         at weblogic.transaction.internal.ServerTransactionImpl.globalPrepare(ServerTransactionImpl.java:2118)
         at weblogic.transaction.internal.ServerTransactionImpl.internalCommit(ServerTransactionImpl.java:259)
         at weblogic.transaction.internal.ServerTransactionImpl.commit(ServerTransactionImpl.java:228)
         at weblogic.transaction.internal.TransactionManagerImpl.commit(TransactionManagerImpl.java:303)
         at weblogic.jms.bridge.internal.MessagingBridge.onMessageInternal(MessagingBridge.java:1279)
         at weblogic.jms.bridge.internal.MessagingBridge.onMessage(MessagingBridge.java:1190)
         at weblogic.jms.adapter.JMSBaseConnection$29.run(JMSBaseConnection.java:1989)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:147)
         at weblogic.jms.adapter.JMSBaseConnection.onMessage(JMSBaseConnection.java:1985)
         at weblogic.jms.client.JMSSession.onMessage(JMSSession.java:2686)
         at weblogic.jms.client.JMSSession.execute(JMSSession.java:2598)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:224)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:183)
    "ExecuteThread: '0' for queue: 'weblogic.kernel.Non-Blocking'":
         at weblogic.transaction.internal.TransactionLoggerImpl$LogDisk.release(TransactionLoggerImpl.java:1322)
         - waiting to lock <a4e99c00> (a weblogic.transaction.internal.TransactionLoggerImpl$IOBuffer)
         at weblogic.transaction.internal.TransactionLoggerImpl.release(TransactionLoggerImpl.java:389)
         at weblogic.transaction.internal.ServerTransactionImpl.releaseLog(ServerTransactionImpl.java:2767)
         at weblogic.transaction.internal.ServerTransactionManagerImpl.remove(ServerTransactionManagerImpl.java:1466)
         at weblogic.transaction.internal.ServerTransactionImpl.setRolledBack(ServerTransactionImpl.java:2597)
         at weblogic.transaction.internal.ServerTransactionImpl.ackRollback(ServerTransactionImpl.java:1093)
         - locked <b23be050> (a weblogic.transaction.internal.ServerTransactionImpl)
         at weblogic.transaction.internal.CoordinatorImpl.ackRollback(CoordinatorImpl.java:298)
         at weblogic.transaction.internal.CoordinatorImpl_WLSkel.invoke(Unknown Source)
         at weblogic.rmi.internal.BasicServerRef.invoke(BasicServerRef.java:492)
         at weblogic.rmi.internal.BasicServerRef$1.run(BasicServerRef.java:435)
         at weblogic.security.acl.internal.AuthenticatedSubject.doAs(AuthenticatedSubject.java:363)
         at weblogic.security.service.SecurityManager.runAs(SecurityManager.java:147)
         at weblogic.rmi.internal.BasicServerRef.handleRequest(BasicServerRef.java:430)
         at weblogic.rmi.internal.BasicExecuteRequest.execute(BasicExecuteRequest.java:35)
         at weblogic.kernel.ExecuteThread.execute(ExecuteThread.java:224)
         at weblogic.kernel.ExecuteThread.run(ExecuteThread.java:183)
    Found 1 deadlock.

  • Bounce file exceeds maximum size. How to change CAF

    I'm new at editing recordings. I only record my speaking engagements. They are rarely under one hour. When I tried to "share" to my iTunes, I got an error message that says
    "The bounce file exceeds the maximum file size. Please change the format to CAF, or decrease the bounce range."
    I have NO idea what that means, and NO idea where I would change the format. Could someone please enlighten me?
    I really appreciate your volunteering your time to share this information!
    ALSO: If I wanted to divide a recording into parts, how would one go about doing this? (maybe making 20 minute sections - and that would decrease the size of the file).
    MANY thanks,
    Lianda 

    When you are sharing to iTunes, try a different quality setting, not AIFF.  A lower quality setting will give a smaller file size.
    To export only a part of the song, enable "Export Cycle Region only" and then use the Cycle region (the yellow bar in the ruler), to mark the part of the song you want to share to iTunes.

  • Mail 5.2(1257) pdf exceeds maximum size

    When attaching a simple PDF (size 15.8MB) to a simple OS Lion email, when going out it expands to 21.4 MB. This exceeds my ISPs 20MB maximun file size. Any suggestion as to why a simple PDF would expand by almost 5 MB.
    The same email without the attachment is only 400 KB.
    Cheers

    PDF is a binary file and as such must be encoded as text before it can be attached to an email. The encoding scheme, base64, adds a third to the size of the original file. What you are seeing is therefore normal and expected. If you want to send a large file to someone, consider putting it on an external server and sending them a link to it instead. There are many such cloud storage services available, many of them free.

  • DBIF_RSQL_INVALID_RSQL The maximum size of an SQL statement was exceeded

    Dear,
    I would appreciate a helping hand
    I have a problem with a dump I could not find any note that I can help solve the problem.
    A dump is appearing at various consultants which indicates the following.
    >>> SELECT * FROM KNA1                     "client specified
    559                  APPENDING TABLE IKNA1
    560                  UP TO RSEUMOD-TBMAXSEL ROWS BYPASSING BUFFER
    ST22
    What happened?
        Error in the ABAP Application Program
        The current ABAP program "/1BCDWB/DBKNA1" had to be terminated because it has
        come across a statement that unfortunately cannot be executed.
    Error analysis
        An exception occurred that is explained in detail below.
        The exception, which is assigned to class 'CX_SY_OPEN_SQL_DB', was not caught
         and
        therefore caused a runtime error.
        The reason for the exception is:
        The SQL statement generated from the SAP Open SQL statement violates a
        restriction imposed by the underlying database system of the ABAP
        system.
        Possible error causes:
         o The maximum size of an SQL statement was exceeded.
         o The statement contains too many input variables.
         o The input data requires more space than is available.
         o ...
        You can generally find details in the system log (SM21) and in the
        developer trace of the relevant work process (ST11).
        In the case of an error, current restrictions are frequently displayed
        in the developer trace.
    SQL sentence
    550     if not %_l_lines is initial.
    551       %_TAB2[] = %_tab2_field[].
    552     endif.
    553   endif.
    554 ENDIF.
    555 CASE ACTION.
    556   WHEN 'ANZE'.
    557 try.
    >>> SELECT * FROM KNA1                     "client specified
    559                  APPENDING TABLE IKNA1
    560                  UP TO RSEUMOD-TBMAXSEL ROWS BYPASSING BUFFER
    561    WHERE KUNNR IN I1
    562    AND   NAME1 IN I2
    563    AND   ANRED IN I3
    564    AND   ERDAT IN I4
    565    AND   ERNAM IN I5
    566    AND   KTOKD IN I6
    567    AND   STCD1 IN I7
    568    AND   VBUND IN I8
    569    AND   J_3GETYP IN I9
    570    AND   J_3GAGDUMI IN I10
    571    AND   KOKRS IN I11.
    572
    573   CATCH CX_SY_DYNAMIC_OSQL_SEMANTICS INTO xref.
    574     IF xref->kernel_errid = 'SAPSQL_ESCAPE_WITH_POOLTABLE'.
    575       message i412(mo).
    576       exit.
    577     ELSE.
    wp trace:
    D  *** ERROR => dySaveDataBindingValue: Abap-Field= >TEXT-SYS< not found [dypbdatab.c  510]
    D  *** ERROR => dySaveDataBindingEntry: dySaveDataBindingValue() Rc=-1 Reference= >TEXT-SYS< [dypbdatab.c  430]
    D  *** ERROR => dySaveDataBinding: dySaveDataBindingEntry() Rc= -1 Reference=>TEXT-SYS< [dypbdatab.c  137]
    Y  *** ERROR => dyPbSaveDataBindingForField: dySaveDataBinding() Rc= 1 [dypropbag.c  641]
    Y  *** ERROR => ... Dynpro-Field= >DISPLAY_SY_SUBRC_TEXT< [dypropbag.c  642]
    Y  *** ERROR => ... Dynpro= >SAPLSTPDA_CARRIER< >0700< [dypropbag.c  643]
    D  *** ERROR => dySaveDataBindingValue: Abap-Field= >TEXT-SYS< not found [dypbdatab.c  510]
    D  *** ERROR => dySaveDataBindingEntry: dySaveDataBindingValue() Rc=-1 Reference= >TEXT-SYS< [dypbdatab.c  430]
    D  *** ERROR => dySaveDataBinding: dySaveDataBindingEntry() Rc= -1 Reference=>TEXT-SYS< [dypbdatab.c  137]
    Y  *** ERROR => dyPbSaveDataBindingForField: dySaveDataBinding() Rc= 1 [dypropbag.c  641]
    Y  *** ERROR => ... Dynpro-Field= >DISPLAY_FREE_VAR_TEXT< [dypropbag.c  642]
    Y  *** ERROR => ... Dynpro= >SAPLSTPDA_CARRIER< >0700< [dypropbag.c  643]
    I thank you in advance
    If you require other information please request

    Hi,
    Under certain conditions, an Open SQL statement with range tables can be reformulated into a FOR ALL ENTRIES statement:
        DESCRIBE TABLE range_tab LINES lines.
        IF lines EQ 0.
          [SELECT for blank range_tab]
        ELSE.
          SELECT .. FOR ALL ENTRIES IN range_tab ..
          WHERE .. f EQ range_tab-LOW ...
          ENDSELECT.
        ENDF.
    Since FOR ALL ENTRIES statements are automatically converted in accordance with the database restrictions, this solution is always met by means of a choice if the following requirements are fulfilled:
    1. The statement operates on transparent tables, on database views or on a projection view on a transparent table.
    2. The requirement on the range table is not negated. Moreover, the range table only contains entries with range_tab-SIGN = 'I'
    and only one value ever occurs in the field range_tab OPTION.
    This value is then used as an operator with operand range_tab-LOW or range_tab-HIGH.In the above example, case 'EQ range_tab-LOW' was the typical case.
    3. Duplicates are removed from the result by FOR ALL ENTRIES.This must not falsify the desired result, that is, the previous Open SQL statement can be written as SELECT DISTINCT.
    For the reformulation, if the range table is empty it must be handled in a different way:with FOR ALL ENTRIES, all the records would be selected here while this applies for the original query only if the WHERE clause consisted of the 'f IN range_tab' condition.
    FOR ALL ENTRIES should also be used if the Open SQL statement contains several range tables.Then (probably) the most extensive of the range tables which fill the second condition is chosen as a FOR ALL ENTRIES table.
    OR
    What you could do in your code is,
    prior to querying;
    since your select options parameter is ultimately an internal range table,
    1. split the select-option values into a group of say 3000 based on your limit,
    2. run your query against each chunck of 3000 parameters,
    3. then put together the results of each chunk.
    For further reading, you might want to have a look at the Note# 13607 as the first suggestion is what I read from the note.

  • Migration Errors: ORA-22973:maximum size allowed

    I am trying to migrate old Content from Portal 3.0 to Oracle9i AS Portal? And am getting this error:
    IMP-00017: following statement failed with ORACLE error 22973:
    "CREATE TABLE "WWSEC_ENABLER_CONFIG_INFO$" OF "SEC_ENABLER_CONFIG_TYPE" OID "
    "'89EE4E7F6D396812E034080020F05106' ( PRIMARY KEY ("LS_LOGIN_URL")) OBJECT "
    "ID PRIMARY KEY PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 LOGGING STORAG"
    "E(INITIAL 131072 NEXT 131072 MINEXTENTS 1 MAXEXTENTS 4096 PCTINCREASE 0 FRE"
    "ELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT) TABLESPACE "USERS""
    IMP-00003: ORACLE error 22973 encountered
    ORA-22973: size of object identifier exceeds maximum size allowed.
    Back in Febuary I have seen posts (Jay and Rich's) that there were some solutions coming down the pike. Are there any solutions or what can be done to solve the error of maximum size allowance?

    What I did was upgrade the initial portal instance with Portal 3.0.6 with the upgrade scripts first. Then import over to a new instance of 3.0.9 with your old content or data. Then, reran the sso_schema script for rebuilding connectivity to the log-in server.
    <BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Carol Kuczborski ([email protected]):
    I just encountered the same exact error trying to export a database from one one machine to another. It looks like all the other portal tables imported OK.
    Have you found a resolution or does anyone know how to successully create the wwsec_enabler_config_info$ table?<HR></BLOCKQUOTE>
    null

  • Credit receipt exceeds Maximum amount allowed

    Hello,
    We were recently requested to update one of our expense type's "Default/Max. value", also we were requested to have the "Amount type" set to "Error message for exceeding maximum".  We have now come across a problem where credit receipts that are uploaded to our system are above the maximum amount allowed and are receiving the error message, these receipts cannot be itemized with personal expense down to the maximum because of the error message, so the users are having to enter these receipts in manually. 
    Is there a way to keep the error message for the maximum value and be able to itemize the receipt even if the receipt is above the limit?
    - Edward Edge

    hi
    no you cannot.
    the only way you can change the system error message as warning and proceed.

  • Ssl certificate exceeds maximum length

    Here is the situation I am having....
    When ever anyone(I have tried, as well as many of my friends and family) tries to log-in to my mail server they get an ssl error saying the the ssl cerificate exceeds maximum permissible length.
    my e-mail web login is page is located here --- https://mail.warezwaldo.us/mail/
    I know that this server is working properly, because I can get to that web page from my Internal Network(with no problems after getting standard cetificate error & excepting the Certificate) but Can Not get to it from the outside world.
    I have verified that SSL is enabled on the Server as well as the Web Browser. I have searched the web for possible solutions to this issue but have yet to find the Solution.
    I am currently on an openSuSE 11.2 Linux Laptop, and have tried in Ubuntu8.04, 9.04, 9.10, and 10.04, as well as Mint7 & 8, and Windows Vista Home Premium & Windows 7, all using FF 3.5.9.
    Can anyone PLEASE HELP, I am trying to start a Business providing Secure E-mail with the ability to have a Web Login Page and this issue is killing me.
    Steps I have taken so far --- I have unistalled and reinstalled the iRedMail system on a server running Ubuntu Server 8.04, 9.04, 9.10, openSuSE server 11.1 & 11.2, CeNTOS 5+, Fedora 10, 11, 12... I have also tried the iRedOS with & with out Updates, and they all give the same error -- ssl certificate exceeds maximum permissible length.
    == URL of affected sites ==
    http://mail.warezwaldo.us/mail/

    OK here is the solution to the Issue that i was having. After dealing with my ISPs crappy equipment I figured out that the issue was being caused by teh Qwest Provided Actiontec PK5000 DSL Modem. Upon initial set-up I had used just the Advanced Port Forwarding and had assigned ports 25, 110, 143, 443, 585, 993, 995 to be forwarded to my mail server, ports 22, 80 to web server, ports 53, 953 to dns server. For 5months this worked just fine and then all of the sudden it stopped working.
    After Dealing Directly with Actiontec Support Staff and being told that I found a "Glitch" in their Software, and well to say the least Actiontec Support Staff couldn't figure out what was causing the Issue and if there was a fix or work around. After about 75hrs of trying to trouble shoot the problem I found the Fix and/or Work Around.
    First: Set-up rules under Advanced Port Forwarding for the appropriate ports to the appropriate IPs
    Second: Create New Rules in Application Forwarding and apply those Newly Created Rules to the Appropriate IPs
    Third: (This is MY RECOMMENDATION) Replace the Actiontec PK5000 ASAP it will stop working on you, this "Glitch" Came after having just Advanced Port Forwarding Rules in place and working fine for 5 months) Third Step to fix the Issue Hope and Pray that the Modem and those Rules will last long enough for you to replace the DSL Modem. I recommend the D-Link DSL2540b it was easy to set-up and runs a lot Cooler than the Actiontec M1000 and PK5000

  • OBI 11g Error : exceeds an entry's the maximum size soft limit 256

    Hi All,
    I am getting a Warning in EM as below:
    Adding property Desc with value "+report description+". exceeds an entry's the maximum size soft limit 256. There are 333 bytes in this property for item /shared/folder/_portal/dashboardname
    It happens only when, if we provide a long description text (>256 bytes) for a report in the 'Description' box while saving the report.
    Do you have any ideas why it is happening and what can be done to remove this warning.
    An parameter needs to be changed..??
    Obi version 11.1.1.5.0
    box : Unix

    We tried that...
    But most of the dashboards/reports have been migrated from 10g and the reports are being built by Users not Dev team, adding their own description to report.
    I need to know ... if there is any parameter which can fix that...

  • Upgrade error  "exceeded maximum allowed length (134217728 bytes)

    Hi:
    I'm trying to add a patch to the OEM patch cache, and I'm getting an error:
    Error: - Failed to Upload File.Uploaded file of length 545451768 bytes exceeded maximum allowed length (134217728 bytes)
    The patch file is what it is (545451768 bytes). How do I install this patch? I'm trying to upgrade a 10.1.0.3 DB to 10.1.0.4
    Thanks

    This log is for external user.
    Did you deploy Lync edge server?
    The Edge Server rejects the authentication request, and redirects the Lync 2010 client to the Lync Web Services (https://lyncexternal.contoso.com/CertProv/CertProvisioningService.svc).
    It seems the redirect fails, please check the event view on Lync edge Server.  
    Please remember to click “Mark as Answer” on the post that helps you, and to click “Unmark as Answer” if a marked post does not actually answer your question. This can be beneficial to other community members reading the thread.

  • ORA-01044: size of buffer bound to variable exceeds maximum

    Hello Oracle Gurus,
    I have a tricky problem.
    I have a stored procedure which has to retun more than 100,000 records. In my stored procedure, I have "TABLE OF VARCHAR2(512) INDEX BY BINARY_INTEGER". It fails when I try to get 80,000 records.
    I get an error "ORA-01044: size 40960000 of buffer bound to variable exceeds maximum 33554432"
    A simple calculation shows that 512*80000=40960000.
    Oracle help suggests to reduce buffer size (i.e., number of records being returned or size of variable).
    But, reducing the number of records returned or reducing the size of variable is not possible because of our product design constraints.
    Are there any other options like changing some database startup parameters to solve this problem?
    Thanks,
    Sridhar

    We are migrating an application running on Oracle 8i to 9i and found the same problem with some of the stored procedures.
    Our setup:
    + Oracle 9.2.0.3.0
    + VB6 Application using OLEDB for Oracle ...
    + MDAC 2.8 msdaora.dll - 2.80.1022.0 (srv03_rtm.030324-2048)
    I am calling a stored procedure from VB like this one:
    {? = call trev.p_planung.GET_ALL_KONTEN(?,?,{resultset 3611, l_konto_id, l_name,l_ro_id, l_beschreibung, l_typ, l_plg_id})}
    If setting the parameter "resultset" beyond a certain limit, I will eventually get this ORA-01044 error. This even happens, if the returned number of records is smaller than what supplied in the resultset parameter (I manually set the "resultset" param in the stored procedure string). E.g.:
    resultset = 1000 -> ORA-06513: PL/SQL: Index der PL/SQL-Tabelle ungültig für Language-Array vom Host
    resultset = 2000 -> OK (actual return: 1043 Recordsets)
    resultset = 3000 -> ORA-01044: Größe 6000000 des Puffers für Variable überschreitet Höchstwert von 4194304
    resultset = 3500 -> ORA-01044: Größe 7000000 des Puffers für Variable überschreitet Höchstwert von 4194304
    ... therefore one record is calculated here 7000000/3500=2000 bytes.
    In Oracle 8i we never had this problem. As this is a huge application using a lot stored procedures, changing all "select" stored procedures to "get data by chunks" (suggestet in some forum threads in OTN) ist not an option.
    Interesting: I can call the stored procedure above with the same parameters as given in VB from e.g. Quest SQL Navigator or sql plus successfully and retrieve all data!
    Is there any other known solution to this problem in Oracle 9i? Is it possible to increase the maximum buffer size (Oracle documentation: ORA-01044 ... Action: Reduce the buffer size.)? What buffer size is meant here - which part in the communication chain supplies this buffer?
    Any help highly appreciated!
    Sincerely,
    Sven Bombach

  • On load, getting error:  Field in data file exceeds maximum length

    Oracle Database 11g Enterprise Edition Release 11.2.0.3.0 - 64bit Production
    PL/SQL Release 11.2.0.3.0 - Production
    CORE    11.2.0.3.0    Production
    TNS for Solaris: Version 11.2.0.3.0 - Production
    NLSRTL Version 11.2.0.3.0 - Production
    I'm trying to load a table, small in size (110 rows, 6 columns).  One of the columns, called NOTES is erroring when I run the load.  It is saying that the column size exceeds max limit.  As you can see here, the table column is set to 4000 Bytes)
    CREATE TABLE NRIS.NRN_REPORT_NOTES
      NOTES_CN      VARCHAR2(40 BYTE)               DEFAULT sys_guid()            NOT NULL,
      REPORT_GROUP  VARCHAR2(100 BYTE)              NOT NULL,
      AREACODE      VARCHAR2(50 BYTE)               NOT NULL,
      ROUND         NUMBER(3)                       NOT NULL,
      NOTES         VARCHAR2(4000 BYTE),
      LAST_UPDATE   TIMESTAMP(6) WITH TIME ZONE     DEFAULT systimestamp          NOT NULL
    TABLESPACE USERS
    RESULT_CACHE (MODE DEFAULT)
    PCTUSED    0
    PCTFREE    10
    INITRANS   1
    MAXTRANS   255
    STORAGE    (
                INITIAL          80K
                NEXT             1M
                MINEXTENTS       1
                MAXEXTENTS       UNLIMITED
                PCTINCREASE      0
                BUFFER_POOL      DEFAULT
                FLASH_CACHE      DEFAULT
                CELL_FLASH_CACHE DEFAULT
    LOGGING
    NOCOMPRESS
    NOCACHE
    NOPARALLEL
    MONITORING;
    I did a little investigating, and it doesn't add up.
    when i run
    select max(lengthb(notes)) from NRIS.NRN_REPORT_NOTES
    I get a return of
    643
    That tells me that the largest size instance of that column is only 643 bytes.  But EVERY insert is failing.
    Here is the loader file header, and first couple of inserts:
    LOAD DATA
    INFILE *
    BADFILE './NRIS.NRN_REPORT_NOTES.BAD'
    DISCARDFILE './NRIS.NRN_REPORT_NOTES.DSC'
    APPEND INTO TABLE NRIS.NRN_REPORT_NOTES
    Fields terminated by ";" Optionally enclosed by '|'
      NOTES_CN,
      REPORT_GROUP,
      AREACODE,
      ROUND NULLIF (ROUND="NULL"),
      NOTES,
      LAST_UPDATE TIMESTAMP WITH TIME ZONE "MM/DD/YYYY HH24:MI:SS.FF9 TZR" NULLIF (LAST_UPDATE="NULL")
    BEGINDATA
    |E2ACF256F01F46A7E0440003BA0F14C2|;|DEMOGRAPHICS|;|A01003|;3;|Demographic results show that 46 percent of visits are made by females.  Among racial and ethnic minorities, the most commonly encountered are Native American (4%) and Hispanic / Latino (2%).  The age distribution shows that the Bitterroot has a relatively small proportion of children under age 16 (14%) in the visiting population.  People over the age of 60 account for about 22% of visits.   Most of the visitation is from the local area.  More than 85% of visits come from people who live within 50 miles.|;07/29/2013 16:09:27.000000000 -06:00
    |E2ACF256F02046A7E0440003BA0F14C2|;|VISIT DESCRIPTION|;|A01003|;3;|Most visits to the Bitterroot are fairly short.  Over half of the visits last less than 3 hours.  The median length of visit to overnight sites is about 43 hours, or about 2 days.  The average Wilderness visit lasts only about 6 hours, although more than half of those visits are shorter than 3 hours long.   Most visits come from people who are fairly frequent visitors.  Over thirty percent are made by people who visit between 40 and 100 times per year.  Another 8 percent of visits are from people who report visiting more than 100 times per year.|;07/29/2013 16:09:27.000000000 -06:00
    |E2ACF256F02146A7E0440003BA0F14C2|;|ACTIVITIES|;|A01003|;3;|The most frequently reported primary activity is hiking/walking (42%), followed by downhill skiing (12%), and hunting (8%).  Over half of the visits report participating in relaxing and viewing scenery.|;07/29/2013 16:09:27.000000000 -06:00
    Here is the full beginning of the loader log, ending after the first row return.  (They ALL say the same error)
    SQL*Loader: Release 10.2.0.4.0 - Production on Thu Aug 22 12:09:07 2013
    Copyright (c) 1982, 2007, Oracle.  All rights reserved.
    Control File:   NRIS.NRN_REPORT_NOTES.ctl
    Data File:      NRIS.NRN_REPORT_NOTES.ctl
      Bad File:     ./NRIS.NRN_REPORT_NOTES.BAD
      Discard File: ./NRIS.NRN_REPORT_NOTES.DSC
    (Allow all discards)
    Number to load: ALL
    Number to skip: 0
    Errors allowed: 50
    Bind array:     64 rows, maximum of 256000 bytes
    Continuation:    none specified
    Path used:      Conventional
    Table NRIS.NRN_REPORT_NOTES, loaded from every logical record.
    Insert option in effect for this table: APPEND
       Column Name                  Position   Len  Term Encl Datatype
    NOTES_CN                            FIRST     *   ;  O(|) CHARACTER
    REPORT_GROUP                         NEXT     *   ;  O(|) CHARACTER
    AREACODE                             NEXT     *   ;  O(|) CHARACTER
    ROUND                                NEXT     *   ;  O(|) CHARACTER
        NULL if ROUND = 0X4e554c4c(character 'NULL')
    NOTES                                NEXT     *   ;  O(|) CHARACTER
    LAST_UPDATE                          NEXT     *   ;  O(|) DATETIME MM/DD/YYYY HH24:MI:SS.FF9 TZR
        NULL if LAST_UPDATE = 0X4e554c4c(character 'NULL')
    Record 1: Rejected - Error on table NRIS.NRN_REPORT_NOTES, column NOTES.
    Field in data file exceeds maximum length...
    I am not seeing why this would be failing.

    HI,
    the problem is delimited data defaults to char(255)..... Very helpful I know.....
    what you need to two is tell sqlldr hat the data is longer than this.
    so change notes to notes char(4000) in you control file and it should work.
    cheers,
    harry

  • Lax validation errors on schema import ('version' exceeds maximum length)

    I have a schema as per below. I'm trying to import it into Oracle 10.2.0.2.0. However, I'm getting the following lax validation error:
    Error loading ora_business_rule.xsd:ORA-30951: Element or attribute at Xpath /schema[@version] exceeds maximum length
    I can fix it by modifying the attribute and shortening it but I'd like to know why it's occuring. Insofar as I can tell there is no imposed limit on the size of schema attributes according to the W3C standard. Which then makes me wonder: does Oracle impose limits on the length of all attributes or is this specific to 'version' ? If there is a limit, what is the upper bound (in bytes) ? Where is this documented?
    Cheers,
    Daniel
    <?xml version="1.0" encoding="UTF-8"?>
    <xs:schema xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns:br="http://foo.com/BusinessRule_PSG_V001" targetNamespace="http://foo.com/BusinessRule_PSG_V001" elementFormDefault="qualified" attributeFormDefault="unqualified" version="last committed on $LastChangedDate: 2006-05-19 11:00:52 +1000 (Fri, 19 May 2006) $">
         <xs:element name="edit">
              <xs:complexType>
                   <xs:sequence>
                        <xs:element name="edit_id" type="xs:string"/>
                        <xs:element ref="br:business_rule"/>
                   </xs:sequence>
              </xs:complexType>
         </xs:element>
         <xs:element name="derivation">
              <xs:complexType>
                   <xs:sequence>
                        <xs:element name="derivation_id" type="xs:string"/>
                        <xs:element ref="br:derivation_type"/>
                        <xs:element ref="br:business_rule"/>
                   </xs:sequence>
              </xs:complexType>
         </xs:element>
         <xs:element name="derivation_type">
              <xs:simpleType>
                   <xs:restriction base="xs:NMTOKENS">
                        <xs:enumeration value="complex"/>
                        <xs:enumeration value="format"/>
                        <xs:enumeration value="formula"/>
                        <xs:enumeration value="recode"/>
                        <xs:enumeration value="SAS code"/>
                        <xs:enumeration value="transfer"/>
                        <xs:enumeration value="count"/>
                        <xs:enumeration value="sum"/>
                        <xs:enumeration value="max"/>
                        <xs:enumeration value="min"/>
                   </xs:restriction>
              </xs:simpleType>
         </xs:element>
         <xs:element name="business_rule"></xs:element>
    </xs:schema>

    Opps -- Sorry it's a decision we took when looking at Version
    When we registered the Schema for Schemas during XDB bootstrap the Version attriubte was mapped to varchar2(12).
    SQL> desc xdb.xdb$schema_T
    Name                                      Null?    Type
    SCHEMA_URL                                         VARCHAR2(700)
    TARGET_NAMESPACE                                   VARCHAR2(2000)
    VERSION                                            VARCHAR2(12)
    NUM_PROPS                                          NUMBER(38)
    FINAL_DEFAULT                                      XDB.XDB$DERIVATIONCHOICE
    BLOCK_DEFAULT                                      XDB.XDB$DERIVATIONCHOICE
    ELEMENT_FORM_DFLT                                  XDB.XDB$FORMCHOICE
    ATTRIBUTE_FORM_DFLT                                XDB.XDB$FORMCHOICE
    ELEMENTS                                           XDB.XDB$XMLTYPE_REF_LIST_T
    SIMPLE_TYPE                                        XDB.XDB$XMLTYPE_REF_LIST_T
    COMPLEX_TYPES                                      XDB.XDB$XMLTYPE_REF_LIST_T
    ATTRIBUTES                                         XDB.XDB$XMLTYPE_REF_LIST_T
    IMPORTS                                            XDB.XDB$IMPORT_LIST_T
    INCLUDES                                           XDB.XDB$INCLUDE_LIST_T
    FLAGS                                              RAW(4)
    SYS_XDBPD$                                         XDB.XDB$RAW_LIST_T
    ANNOTATIONS                                        XDB.XDB$ANNOTATION_LIST_T
    MAP_TO_NCHAR                                       RAW(1)
    MAP_TO_LOB                                         RAW(1)
    GROUPS                                             XDB.XDB$XMLTYPE_REF_LIST_T
    ATTRGROUPS                                         XDB.XDB$XMLTYPE_REF_LIST_T
    ID                                                 VARCHAR2(256)
    VARRAY_AS_TAB                                      RAW(1)
    SCHEMA_OWNER                                       VARCHAR2(30)
    NOTATIONS                                          XDB.XDB$NOTATION_LIST_T
    LANG                                               VARCHAR2(4000)
    SQL>

  • Incoming message size exceeds the configured maximum size for protocol t3

    Hi All,
    I've encountered an error as follow:
    weblogic.socket.MaxMessageSizeExceededException: Incoming message of size 50004000 bytes exceeds the configured maximum of 50000000 bytes of protocol t3.
    But the request message is only 3MB, why it is enlarged to over 50M?
    There is a For Each loop section in main flow, is it because for one loop, there will be a copy of request message?
    How to enlarge message size for protocol t3?
    Go to server/protocol and change 'Maximum Message Size' for AdminServer, OSB Servers and SOA servers?
    Thanks and Regards,
    Bruce

    Hi,
    1) After setting -Dweblogic.MaxMessageSize to 25000000
    <BEA-000403> <IOException occurred on socket: Socket[addr=ac-sync-webserver1/172.24.128.8,port=9040,localport=36285]
    weblogic.socket.MaxMessageSizeExceededException: Incoming message of size: '25002240' bytes exceeds the configured maximum of: '25000000' bytes for protocol: 't3'
    at weblogic.socket.BaseAbstractMuxableSocket.incrementBufferOffset(BaseAbstractMuxableSocket.java:174)
    2) After setting -Dweblogic.MaxMessageSize to 50000000
    <BEA-000403> <IOException occurred on socket: Socket[addr=ac-sync-webserver2/172.24.128.9,port=9040,localport=59925]
    weblogic.socket.MaxMessageSizeExceededException: Incoming message of size: '50000400' bytes exceeds the configured maximum of: '50000000' bytes for protocol:
    't3'.
    And even after setting various values for -Dweblogic.MaxMessageSize , issue weblogic.socket.MaxMessageSizeExceededException was observed.
    To overcome the issue set Manual Service Migration Only as after several experiments and replicating the issue it was found out that in case of no available pinned services, must set the migration policies of the migratable targets on "Manual Service Migration Only".
    And once it is corrected; it was noticed that weblogic.socket.MaxMessageSizeExceededException issue also resolved.
    WebLogic Server can fail over most services transparently, but it's unable to do the same when dealing with pinned services.
    Pinned Services : JMS and JTA are considered as pinned services. They're hosted on individual members of a cluster and not on all server instances.
    You can have high availability only if the cluster can ensure that these pinned services are always running somewhere in the cluster.
    When a WebLogic Server instance hosting these critical pinned services fails, WebLogic Server can't support their continuous availability and uses migration instead of failover to ensure that they are always available.
    Regards,
    Kal

Maybe you are looking for

  • Custom program problem

    Moderator message - please use meaningful subject in future Hi, My requirement is : We get a input file in .XLS format. To load, it has to be converted to .TXT format. But this conversion sometimes creates junk characters in the load file. so to rect

  • Putting images back into iPhoto 4.0.3 - why so many?

    Hi A while ago made some back-up DVDs of images - 2004, 2005 and so on - using iP to do so. Then we removed all but 2007's pics from iP to make more room, although by your standards we had hardly anything in the library (at the time an iMovie project

  • Report server, error code 186

    I installed 9ias(http server, forms &report server , oem server) on windows 2000, but the report server don't startup, error code 186. can you tell me how to resolve the problem.

  • My Zip files aren't unzipping

    I receive PDF-zipped files for work via email attachments, but when I double click on them, the window disappears (a brief flash of a message saying the file is being "archived" flashes, but then nothing.) When I go to my user>library>maildownloads,

  • How to fix Waiting on time service?

    How do you repair the "waiting on time service" failure.  I have tried the NI support page solution to restart the time service.  It did not work.  It suggested downloading a later version of logos.  The available version stated it was older than the