Script Command limit exceeded! - incorrect

I have a project which I have a lot of scripts in,
every time I try and build it it says Command limit exceeded! and gives me the name of the script.
however I am way under the limit and even if i delete all the lines of code in that script and then try and built it still shows the same error. I haven't done any GPRM partitioning so I can't see what i am doing wrong. Even if i delete that script entirely another one will be flagged up in the same way, even if it only has 3 lines in it.
I have tryed it on 3 machines with DSP 4.2.1 and 4.2.2
any help greatly appreciated
al

I would think the project file is somehow corrupted. You certainly won't be up against any limits if all your scripts are this short, AFAIK.
Have you tried saving this out with a different project name and re-opening that? Deleting the PAR files and so on... all the usual remedies?
There is also a way to script the same thing but use less scripts (but more complex math). You can go to any menu you want if you know the value of the menu (menu 1 has a value of 32, menu 2 is 64 and so on). Using some fairly simple calculations you can get to just about any location you want with a few lines of code. You'll be under the command limit, but the scripting is a little more obscure than your current way.
So, filling a GPRM with a value of 32, and then using a 'Jump Indirect' will take you to the first menu:
Mov GPRM0, 32
Jump Indirect GPRM0
You can do this for tracks and stories too, but the values are different. Track 1 has a value of 49280. They increment in jumps of 128, and include stories too - so track 2, story 1 would be the equivalent of 3x128+49280.
The thing is, it doesn't matter which track you have got - this all relates to the layout in the outline view. Track 1 is always the one at the top of the list - and has the value 49280. You can move content around in the list and still know that 49280 will get you to track 1. The same is true for your menus - whichever is at the top of the list is menu 1, and a jump indirect of 32 will get you to it.
What this will do is reduce the total number of scripts you are using and reduce the number of commands in use. It will also give you much more flexibility as you never have to use menu names and track names in your scripts which is useful if you are re-using projects and altering the number of items in each. However, it won't be as simple to follow through as the extended scripting you are using now... so much depends on where you are with your understanding of scripts, I guess!
Have a look here for some more general info:
http://www.dvdstudiopro.co.uk/tutorials/play-all-3/
Remember, the files in that tutorial are demo files... you'll need to set your disc up more thoroughly that I've done there if you want to use the sample as a real production.

Similar Messages

  • Mails to 1 particular user is getting bounced - Command time limit exceeded

    I've about 100 mail clients. Since yesterday morning, I've had one user whose incoming emails have been getting bounced. I've run a "mailbfr -m user" w/o success. The server is current on updates and has been restarted. All other users are functioning correctly. Notes:
    -IMAP account
    -This user has a 4GB Sent folder. All other folders are under 2GB.
    -This user has recently starting syncing Notes from a Blackberry.
    THE SMTP LOG:
    Oct 7 08:26:52 myserver postfix/pipe[488]: D1BCE11CE230: to=<[email protected]>, relay=cyrus, delay=1000, delays=0.01/0/0/1000, dsn=5.3.0, status=bounced (Command time limit exceeded: "/usr/bin/cyrus/bin/deliver")
    A PORTION OF THE BOUNCE EMAIL SENT TO SENDER:
    This is the mail system at host myserver.com.
    I'm sorry to have to inform you that your message could not
    be delivered to one or more recipients. It's attached below.
    For further assistance, please send mail to postmaster.
    If you do so, please include this problem report. You can
    delete your own text from the attached returned message.
    The mail system
    <[email protected]>: Command time limit exceeded:
    "/usr/bin/cyrus/bin/deliver"
    Reporting-MTA: dns; myserver.com
    X-Postfix-Queue-ID: AD79711CE1DD
    X-Postfix-Sender: rfc822; [email protected]
    Arrival-Date: Thu, 7 Oct 2010 08:07:57 -0700 (PDT)
    Final-Recipient: rfc822; [email protected]
    Original-Recipient: rfc822;[email protected]
    Action: failed
    Status: 5.3.0
    Diagnostic-Code: x-unix; internal software error
    Any help?

    Solved. This person was in-process of moving Blackberry notes to the imap account. There was an interruption during the transfer phase and the server's Notes folder became bad somehow. When Mailbfr hit that particular folder, it stopped the rebuild process.
    I removed the Notes folder and ran Mailbfr again. The Notes folder was recreated and the rebuild function completed.
    Mail delivery has been restored to this user.

  • PI 7.11 - SXI_CACHE Issue - SQL 0904 - Resource Limit Exceeded

    Hi PI & IBM I Gurus,
    We are having the SXI_CACHE issue in our PI 7.11 SPS04 system. When we try to do the delta cache refresh through SXI_CACHE, it is returning the error SQL 0904 - Resource limit exceeded. When we try to do the full cache refresh, it is getting the issue 'Application issue during request processing'.
    We have cleaned up the SQL Packages with DLTR3PKG command, which did not resolve the issue. We recently performed a system copy to build the QA instance and I observed that the adapter engine cache for the development is presented in the QA instance and removed that cache from there.
    I am not seeing the adapter engine connection data cache in our PI system. The adapter engine cache is working fine.
    All the caches are working fine from the PI Administration page. The cache connectivity test is failing with the same error as I mentioned for the SXI_CACHE.
    Please let me know if you have encountered any issue like this on IBM I 6.1 Platform.
    Your help is highly appreciated.
    Thanks
    Kalyan

    Hi Kalyan,
    SQL0904 has different reason codes ...
    Which one are you seeing ?
    Is the SQL pack really at its boundary of 1GB ?
    ... otherwise, it is perhaps a totally different issue ... then DLTR3PKG cannot help at all ...
    If you should see this big SQL Package, you should use PRTSQLINF in order to see if there is more or less over and over the same SQL in, just with different host variables or so ...
    If the last point should be the case, I would open a message with BC-DB-DB4 so that they can check how to help here or to talk to the application people to behave a bit different ...
    Regards
    Volker Gueldenpfennig, consolut international ag
    http://www.consolut.com http://www.4soi.de http://www.easymarketplace.de

  • PI 7.11 - SXI_CACHE Issue - SQL0904 - Resource limit exceeded

    Hi IBM I Gurus,
    We are having the SXI_CACHE issue in our PI 7.11 SPS04 system. When we try to do the delta cache refresh through SXI_CACHE, it is returning the error SQL 0904 - Resource limit exceeded. When we try to do the full cache refresh, it is getting the issue 'Application issue during request processing'.
    We have cleaned up the SQL Packages with DLTR3PKG command, which did not resolve the issue. We recently performed a system copy to build the QA instance and I observed that the adapter engine cache for the development is presented in the QA instance and removed that cache from there.
    I am not seeing the adapter engine connection data cache in our PI system. The adapter engine cache is working fine.
    All the caches are working fine from the PI Administration page. The cache connectivity test is failing with the same error as I mentioned for the SXI_CACHE.
    Please let me know if you have encountered any issue like this on IBM I 6.1 Platform.
    Your help is highly appreciated.
    Thanks
    Kalyan

    Hi Kalyan,
    SQL0904 has different reason codes ...
    Which one are you seeing ?
    Is the SQL pack really at its boundary of 1GB ?
    ... otherwise, it is perhaps a totally different issue ... then DLTR3PKG cannot help at all ...
    If you should see this big SQL Package, you should use PRTSQLINF in order to see if there is more or less over and over the same SQL in, just with different host variables or so ...
    If the last point should be the case, I would open a message with BC-DB-DB4 so that they can check how to help here or to talk to the application people to behave a bit different ...
    Regards
    Volker Gueldenpfennig, consolut international ag
    http://www.consolut.com http://www.4soi.de http://www.easymarketplace.de

  • Please HELP with ORA-19809: limit exceeded for recovery files

    Hi Gurus,
    I am not a DBA but a developer. Today my Oracle database 11g on my PC will not open and I am getting the errors,
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    I checked my alert logs and found these errors,
    Errors in file c:\oracle\diag\rdbms\orcl\orcl\trace\orcl_ora_2144.trc:
    ORA-19809: limit exceeded for recovery files
    ORA-19804: cannot reclaim 41072128 bytes disk space from 4039114752 limit
    ARCH: Error 19809 Creating archive log file to 'C:\ORACLE\FLASH_RECOVERY_AREA\ORCL\ARCHIVELOG\2011_04_08\O1_MF_1_120_%U_.ARC'
    Errors in file c:\oracle\diag\rdbms\orcl\orcl\trace\orcl_ora_2144.trc:
    ORA-16038: log 3 sequence# 120 cannot be archived
    ORA-19809: limit exceeded for recovery files
    ORA-00312: online log 3 thread 1: 'C:\ORACLE\ORADATA\ORCL\REDO03.LOG'
    USER (ospid: 2144): terminating the instance due to error 16038
    Instance terminated by USER, pid = 2144
    So I started searching for answers on this forum and found a suggestion using RMAN to delete all the archive logs,
    RMAN> connect target sys/*******
    RMAN> delete archivelog all;
    RMAN>crosscheck archivelog all;
    But I get these errors in RMAN,
    using target database control file instead of recovery catalog
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03002: failure of delete command at 04/08/2011 11:20:58
    RMAN-06403: could not obtain a fully authorized session
    ORA-01034: ORACLE not available
    ORA-27101: shared memory realm does not exist
    So, please help me: How do I fix this problem since I can't start the instance?
    Thank you.
    Ben

    You need to increase the FRA by changing the parameter DB_RECOVERY_FILE_DEST_SIZE. Create a parameter file while being connected as Sys user and issue,
    sql>create pfile from spfile;It would create the pfile in the Oracle_Home/database folder. Edit that file and increase the parameter to a larger size-like double of it's current value. Save the file and issue,
    sql>startup pfile=<your location of the file>HTH
    Aman....

  • Java Session Space Limit Exceeded warning

    Hi,
    This is related to runtime memory. I have java programs which have recursive calls and also it is highly data intensive. The problem is this.
    When i loaded the classes inside oracle database (using loadjava command) and run the program inside oracle i got session space limit exceeded warning. So i increased the default value of 1MB session space limit to 5 MB. When i ran the program i got the error again. I tried again with 20MB, 40B still i got the error.
    In my next step to solve this problem, i ran the program outside oracle and started a MemoryWatcher program along with it to see how much is consumed at the runtime. I printed Runtime.totalMemory() and Runtime.freeMemory() in the MemoryWatcher thread.
    I am able to see the Total memory creeping up from 2MB to 2.5MB, 3MB.....
    But the free memory varies between 0.3MB to 1.5MB when the program is running.
    What could be the reason the Java session space limit exceeds even when the space limit parameter is set to 40MB.
    How can i avoid this warning??. Any help on this would be greatly appreciated.
    Thanks in advance!!. Pls send your response as a copy to [email protected]
    Regards,
    Babu

    Oracle Team, nobody knows the solutions???
    Or is bug.
    Best regards, Marcelo.
    Oracle 8.1.7 (sun/linux) and XDK 9.0.2+ parsing big documents.

  • ORA-19809: limit exceeded for recovery files and flash_recovery_area

    Hi.
    I am getting the below errors (at the end) when trying to backup the database using RMAN.
    The directory structure is
    E:\oracle\TEST
    E:\oracle\TEST\flash_recovery_area
    E:\oracle\TEST\flash_recovery_area\arch --> HAS ARCHIVE FILES
    E:\oracle\TEST\flash_recovery_area\TEST
    E:\oracle\ATEST\flash_recovery_area\TEST\BACKUPSET --> RMAN backup sets
    E:\oracle\TEST\flash_recovery_area\TEST\ONLINELOG --> This is empty.
    D:\oracle\product\10.2.0\ is oracle home
    It has:
    D:\oracle\product\10.2.0\admin
    D:\oracle\product\10.2.0\db1
    D:\oracle\product\10.2.0\oradata
    D:\oracle\product\10.2.0\oradata\test has data files.
    1) I created this database using OUI and DBCA and I have a feeling that this structure is somewhat convoluted. Could you let me know if this needs to be changed?
    2) Secondly, the RMAN error. The entire size of E:\oracle\TEST\flash_recovery_area is now only 1.74 GB yet I am getting this error.
    I deleted old archive logs from E:\oracle\TEST\flash_recovery_area\arch . Currently it has only arch logs for seven days. Next I issued delete expired archivelog all command. Right now crosscheck archivelog all finds only 40 archive logs..
    and increased db_recovery_file_dest_size to 10GB. alter system set db_recovery_file_dest_size=10000M scope=both;
    select name
    , floor(space_limit / 1024 / 1024) "Size MB"
    , ceil(space_used / 1024 / 1024) "Used MB"
    from v$recovery_file_dest
    order by name
    This query is returning the use space as 6GB. However physically only 1.74 GB is currenly used. Please let me know the way forward. Appreciate in advance.
    channel ORA_DISK_1: starting piece 1 at 09-JUN-09
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 06/09/2009 17:56:
    15
    ORA-19809: limit exceeded for recovery files
    ORA-19804: cannot reclaim 52428800 bytes disk space from 10485760000 limit
    continuing other job steps, job failed will not be re-run
    channel ORA_DISK_1: starting full datafile backupset
    channel ORA_DISK_1: specifying datafile(s) in backupset
    including current control file in backupset
    including current SPFILE in backupset
    channel ORA_DISK_1: starting piece 1 at 09-JUN-09
    channel ORA_DISK_1: finished piece 1 at 09-JUN-09
    piece handle=E:\ORACLE\TEST\FLASH_RECOVERY_AREA\TEST\BACKUPSET\2009_06_09\
    O1_MF_NCSNF_TAG20090609T175224_52XPN0Y9_.BKP tag=TAG20090609T175224 comment=NONE
    channel ORA_DISK_1: backup set complete, elapsed time: 00:00:02
    RMAN-00571: ===========================================================
    RMAN-00569: =============== ERROR MESSAGE STACK FOLLOWS ===============
    RMAN-00571: ===========================================================
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 06/09/2009 17:56:
    15
    ORA-19809: limit exceeded for recovery files
    ORA-19804: cannot reclaim 52428800 bytes disk space from 10485760000 limit
    RMAN-03009: failure of backup command on ORA_DISK_1 channel at 06/09/2009 17:54:
    19
    ORA-19809: limit exceeded for recovery files
    ORA-19804: cannot reclaim 52428800 bytes disk space from 10485760000 limit
    RMAN>

    Hi,
    Thanks for looking into my problem... Here are the answers..
    SQL> archive log list;
    Database log mode Archive Mode
    Automatic archival Enabled
    Archive destination E:\oracle\TEST\flash_recovery_area\arch
    Oldest online log sequence 1011
    Next log sequence to archive 1013
    Current log sequence 1013
    SQL>
    select * from v$flash_recovery_area_usage;
    FILE_TYPE     PERCENT_SPACE_USED     PERCENT_SPACE_RECLAIMABLE     NUMBER_OF_FILES
    CONTROLFILE     0.00               0.00                    0
    ONLINELOG     0.00               0.00                    0
    ARCHIVELOG     0.00               0.00                    0
    BACKUPPIECE     66.57               0.20                    5
    IMAGECOPY     0.00               0.00                    0
    FLASHBACKLOG     0.00               0.00                    0
    SQL> show parameter log_archive_dest
    NAME TYPE VALUE
    log_archive_dest string
    log_archive_dest_1 string LOCATION=E:\oracle\TEST\flash_recovery_area\arch
    log_archive_dest_10 string
    log_archive_dest_2 string
    log_archive_dest_3 string
    log_archive_dest_4 string
    log_archive_dest_5 string
    log_archive_dest_6 string
    log_archive_dest_7 string
    log_archive_dest_8 string
    NAME TYPE VALUE
    log_archive_dest_9 string
    log_archive_dest_state_1 string enable
    log_archive_dest_state_10 string enable
    log_archive_dest_state_2 string enable
    log_archive_dest_state_3 string enable
    log_archive_dest_state_4 string enable
    log_archive_dest_state_5 string enable
    log_archive_dest_state_6 string enable
    log_archive_dest_state_7 string enable
    log_archive_dest_state_8 string enable
    log_archive_dest_state_9 string enable
    SQL>

  • Administrative limit exceeded error

    On attempting an ldap search command as follows ./ldapsearch -z1000 -b "ou=People,o=abcd.com" "organizationalstatus=manager", I am getting "Administrative limit exceeded" error. The limit is set to 2000 in the console. any help is appreciated. Thanks

    hello,
    "Administrative limit " may be a SIZE limit , ENTRIES ( number of ) limit or SIZE ( output of ) limit:
    From "man" of ldapsearch:
    -l timelim time limit (in seconds) for search (default is no limit)
    -z sizelim size limit (in entries) for search (default is no limit)
    You can use these options when you search.
    You may also check the setting of 3 limits above with Admin Console:
    Open your directory server (name), Configuration Folder, Performance. On right panel check Client Control folder. You may see:
    Size limit
    Look-through limit
    Time limit
    Idle timeout
    � The look-through limit specifies the maximum number of entries that will be examined for a search operation.
    � The size limit specifies the maximum number of entries the server returns to the client application in response to a search operation.
    � The time limit specifies the maximum time the server spends processing a search operation.
    � The idle timeout specifies the time a client connection to the server can be idle before the server drops the connection.
    If you bind as Directory manager, you may use unlimited resources by default.
    Hello,
    silvio

  • Message limit exceeded , remaining 3 errors not reported

    hi,
    Whenever I hit review form in ISR , I am getting a popup with this message , Message limit exceeded , remaining 3 errors not reported. Can anybody where I can see the error.
    Thanks,
    Manish

    This could probably be a script error (warning), which you can find in the report pallette.
    Else it could be that you have entered text of different font , which is not supported by adobe. (like say you have copied and pasted a fancy text from an html page).
    If it is from html or xml pages, chances can be that the backend codes interfere and results in the error.
    We faced similar issues, that was coz of  usage of 'fancy' fonts.
    Thanks and Regards,
    Antony John Isacc.

  • ReadString() get time limit exceed to complete operation error

    I have a device name Time Electronics 5075. I am trying to use the SimpleReadWrite.2010 in National Instruments\NI-488.2\Examples\DotNET4.0\SimpleReadWrite\cs to comunicate with the device.
    I can open the connection and send commands to the device by using this code
    device = new Device((int)boardIdNumericUpDown.Value,(byte)primaryAddressNumericUpDown.Value,(byte)currentSecondaryAddress);
    device.IOTimeout = TimeoutValue.T100s;
    device.Write("X0/A1/F0{13}");
     But I am stuck at 
    stringReadTextBox.Text = device.ReadString();
    I get the error "Time limit exceeded to complete operation."
    I don't know exactly should I add 
    device.Write("T/D{13}");
    before the device.ReadString() or not. But I get the same error for both cases.
    Some info: Visual Studio 2010, Measurement Studio 2013, NI.488.2, Windows 7 Pro
    Any idea?

    Thank you Curt_C,
    I have attached the screen capture of NI IO Trace. I just open the session, send two commands T\n and D\n to the device and readstring. The red one are errors on reading. The last one is close session.
    I think \n is the correct terminator. It works with most of commands, just get stuck with ReadString.
    Kind regards,
    Thang
    PS: I also attached GPIB Interface Properties for more info
    Attachments:
    NI IO Trace_01.png ‏54 KB
    GPIB Interface Properties.png ‏54 KB

  • RFC_ERROR_SYSTEM_FAILURE: Time limit exceeded. Connection Pool - JCO api

    Hi Everyone
    My Connection  Pool parameters JCO api.
    client=300
    user=SISGERAL_RFC
    passwd=******
    ashost=14.29.3.120
    sysnr=00
    size=10
    I have these parameters on my Connection Pool and sometimes appear these wrongs in my application:
    1.
    2006-01-07 13:20:37,414 ERROR com.tel.webapp.framework.SAPDataSource - ##### Time limit exceeded. LOCALIZED MESSAGE = Time limit exceeded. KEY = RFC_ERROR_SYSTEM_FAILURE GROUP = 104 TOSTRING = com.sap.mw.jco.JCO$Exception: (104) RFC_ERROR_SYSTEM_FAILURE: Time limit exceeded.
    2.
    2006-01-07 14:01:31,007 ERROR com.tel.webapp.framework.SapPoolConnectionManager - Timeout
    I’d like to know if is happening.
    Are there something wrong with my connection pool?
    What can be happening?
    Thanks

    Raghu,
    Thanks for your response.
    Yes, the pool connections are in place according to the sAP note mentioned above.
    Regards,
    Faisal

  • Memory Leak, Receiver Got Null Message & Consumer limit exceeded on destina

    When running program that adds an Object message to a JMS queue and then recieves it. I get the following.
    1) interminitent NULL messages recieved.
    2) jms.JMSException: [C4073]: Consumer limit exceeded on destination interactionQueueDest. Even though only one receiver can be receiving via the supplied program.
    3) After many message are added to the queue 1000's the Message Queue goes to Out Of Memory exception. It should swap to disk!!
    STEPS TO FOLLOW TO REPRODUCE THE PROBLEM :
    RUN this program via a JSP call in the application server.
    JSP
    <%@ page language="java" import="jms.*"%>
    <html>
    <head>
    <title>Leak Memory</title>
    </head>
    <body>
    <hr/>
    <h1>Leak Memory</h1>
    <%
       LeakMemory leakMemory = new LeakMemory();
       leakMemory.runTest(10000,1000);
      // NOTE will brake but slower with setting leakMemory.runTest(10000,100);
    %>JMS resources must be created:
    jms/queueConnectionFactory
    jms/interactionQueue
    must be created first.
    Class:
    package jms;
    import javax.naming.*;
    import javax.jms.*;
    public class LeakMemory implements Runnable {
      private QueueConnectionFactory queueConnectionFactory = null;
      private Queue interactionQueue = null;
      private boolean receiverRun = true;
      private QueueConnection queueConnection;
      private int totalMessageInQueue = 0;
      public LeakMemory() {
        init();
       * initialize queues
      public void init(){
        try {
          InitialContext context = new InitialContext();
          this.queueConnectionFactory = (QueueConnectionFactory)context.lookup("jms/queueConnectionFactory");
          this.interactionQueue = (Queue) context.lookup("jms/interactionQueue");
        catch (NamingException ex) {
          printerError(ex);
      public void runTest(int messageCount, int messageSize){
        this.receiverRun = true;
        Thread receiverThread = new Thread(this);
        receiverThread.start();
        for (int i = 0; i < messageCount; i++) {
            StringBuffer messageToSend = new StringBuffer();
            for (int ii = 0; ii < messageSize; ii++) {
              messageToSend.append("0123456789012345678901234567890123456789012345678901234567890123456789012345678901234567890123456789\n");
            QueueConnection queueConnectionAdder = null;
            QueueSession queueInteractionSession = null;
            QueueSender interactionQueueSender = null;
            try {
                //Get a queue connection
                queueConnectionAdder = this.getQueueConnection();
                queueInteractionSession = queueConnectionAdder.createQueueSession(false, Session.AUTO_ACKNOWLEDGE);
                interactionQueueSender = queueInteractionSession.createSender(this.interactionQueue);
                ObjectMessage objectMessage = queueInteractionSession.createObjectMessage(messageToSend);
                objectMessage.setStringProperty("PROPERTY", "" + System.currentTimeMillis());
                //Send object
                interactionQueueSender.send(objectMessage,DeliveryMode.PERSISTENT,5,0);
                totalMessageInQueue++;
                //Close Resources
                interactionQueueSender.close();
                queueInteractionSession.close();
                interactionQueueSender = null;
                queueInteractionSession = null;
            } catch (JMSException ex) {
               printerError(ex);
       * run
      public void run()  {
        while(this.receiverRun){
                try {
                  QueueSession interactionQueueSession = this.getQueueConnection().createQueueSession(false, Session.CLIENT_ACKNOWLEDGE);
                  QueueReceiver queueReceiver = interactionQueueSession.createReceiver(this.interactionQueue);
                  ObjectMessage message = (ObjectMessage)queueReceiver.receive(100);
                  if(message != null){
                    StringBuffer messageRecived = (StringBuffer)message.getObject();
                    //Simulate Doing Something
                    synchronized (this) {
                      try {
                        Thread.sleep(400);
                      catch (InterruptedException ex1) {
                        //Can Safely be ignored
                    message.acknowledge();
                    totalMessageInQueue--;
                  } else {
                    printerError(new Exception("Receiver Got Null Message"));
                  queueReceiver.close();
                  interactionQueueSession.close();
                  queueReceiver = null;
                  interactionQueueSession = null;
                catch (JMSException ex) {
                  printerError(ex);
        * Get's the queue Connection and starts it
        * @return QueueConnection The queueConnection
       public synchronized QueueConnection getQueueConnection(){
           if (this.queueConnection == null) {
             try {
               this.queueConnection = this.queueConnectionFactory.createQueueConnection();
               this.queueConnection.start();
             catch (JMSException ex) {
               printerError(ex);
         return this.queueConnection;
       private void printerError(Exception ex){
         System.err.print("ERROR Exception totalMessageInQueue = " + this.totalMessageInQueue + "\n");
         ex.printStackTrace();
    }Is there something wrong with the way I'm working with JMS or is it just this unreliable in Sun App Server 7 Update 3 on windows?

    1) interminitent NULL messages recieved.Thanks that explains the behavior. It was wierd getting null messages when I know there is messages in the queue.
    2) jms.JMSException: [C4073]: Consumer limit exceeded
    on destination interactionQueueDest. Even though only
    one receiver can be receiving via the supplied
    program. No other instances, Only this program. Try it yourself!! It works everytime on Sun Application Server 7 update 2 & 3
    heres the broker dump at that error point
    [14/Apr/2004:12:51:47 BST] [B1065]: Accepting: [email protected]:3211->admin:3205. Count=1
    [14/Apr/2004:12:51:47 BST] [B1066]:   Closing: [email protected]:3211->admin:3205. Count=0
    [14/Apr/2004:12:52:20 BST] [B1065]: Accepting: [email protected]:3231->jms:3204. Count=1 [14/Apr/2004:12:53:31 BST] WARNING [B2009]: Creation of consumer  from connection [email protected]:3231 on destination interactionQueueDest failed:
    B4006: com.sun.messaging.jmq.jmsserver.util.BrokerException: [B4006]: Unable to attach to queue queue:single:interactionQueueDest: a primary queue is already active
    3) After many message are added to the queue 1000's
    the Message Queue goes to Out Of Memory exception. It
    should swap to disk!!The broker runs out of memory. Version in use
    Sun ONE Message Queue                   Copyright 2002
    Version:  3.0.1 SP2 (Build 4-a)              Sun Microsystems, Inc.
    Compile:  Fri 07/11/2003         All Rights ReservedOut of memory snippet
    [14/Apr/2004:13:08:28 BST] [B1089]: In low memory condition, Broker is attempting to free up resources
    [14/Apr/2004:13:08:28 BST] [B1088]: Entering Memory State [B0022]: YELLOW from previous state [B0021]: GREEN  - current memory is 118657K, 60% of total memory
    [14/Apr/2004:13:08:38 BST] WARNING [B2075]: Broker ran out of memory before the passed in VM maximum (-Xmx) 201326592 b,  lowering max to currently allocated memory (200431976 b ) and trying to recover [14/Apr/2004:13:08:38 BST] [B1089]: In low memory condition, Broker is attempting to free up resources
    [14/Apr/2004:13:08:38 BST] [B1088]: Entering Memory State [B0024]:  RED  from previous state [B0022]: YELLOW - current memory is 128796K, 99% of total memory [14/Apr/2004:13:08:38 BST] ERROR [B3008]: Message 2073-192.168.0.50(80:d:b6:c4:d6:73)-3319-1081944517772 exists in the store already [14/Apr/2004:13:08:38 BST] WARNING [B2011]: Storing of JMS message from IMQConn[AUTHENTICATED,[email protected]:3319,jms:3282] failed:
    com.sun.messaging.jmq.jmsserver.util.BrokerException: Message 2073-192.168.0.50(80:d:b6:c4:d6:73)-3319-1081944517772 exists in the store already
    [14/Apr/2004:13:08:38 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:38 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:41 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:44 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:44 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:44 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:45 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:45 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:46 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:46 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:48 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:52 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:52 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:53 BST] ERROR [B3107]: Attempt to free memory failed, taking more drastic measures : java.lang.OutOfMemoryError
    [14/Apr/2004:13:08:53 BST] ERROR unable to deal w/ error: 1
    [14/Apr/2004:13:08:53 BST] ERROR TRYING TO CLOSE [14/Apr/2004:13:08:53 BST] ERROR DONE CLOSING
    [14/Apr/2004:13:08:53 BST] [B1066]:   Closing: [email protected]:3319->jms:3282. Count=0

  • Short dump "Time limit exceeded" when searching for Business Transactions

    Hello Experts,
    We migrated from SAP CRM 5.2 to SAP CRM 7.0. After migration, our business transaction search (quotation, sales order, service order, contract etc) ends with the short dump "Time limit exceeded" in class CL_CRM_REPORT_ACC_DYNAMIC, method DATABASE_ACCESS. The select query is triggered from line 5 of this method.
    Number of Records:
    CRMD_ORDERADM_H: 5,115,675
    CRMD_ORDER_INDEX: 74,615,914
    We have done these so far, but the performance is still either poor or times out.
    1. DB team checked the ORACLE parameters and confirmed they are fine. They also checked the health of indices in table CRMD_ORDER_INDEX and indices are healthy
    2. Created additional indices on CRMD_ORDERADM_H and CRMD_ORDER_INDEX. After the creation of indices, some of the searches(without any criteria) work. But it takes more than a minute to fetch 1 or 2 records
    3. An ST05 trace confirmed that the selection on CRMD_ORDER_INDEX takes the most time. It takes about 103 seconds to fetch 2 records (max hits + 1)
    4. If we specify search parameters, say for example a date or status, then again we get a short dump with the message "Time limit exceeded".
    5. Observed that only if a matching index is available for the WHERE clause, the results are returned (albeit slowly). In the absence of an index, we get the dump.
    6. Searched for notes and there are no notes that could help us.
    Any idea what is causing this issue and what we can do to resolve this?
    Regards,
    Bala

    Hi Michael,
    Thanks. Yes we considered the note 1527039. None of the three scenarios mentioned in the note helped us. But we ran CRM_INDEX_REBUILD to check if the table CRMD_ORDER_INDEX had a problem. That did not help us either.
    The business users told us that they mostly search using the date fields or Object ID. We did not have any problem with search by Object ID. So we created additional indices to support search using the date fields.
    Regards,
    Bala

  • Static credit check: credit limit exceeded for consignment issue delivery

    Hi,
    We are encountering a credit limit exceeded error in delivery creation for a consignment issue.
    In SPRO config "Credit limit check for order types", no credit limit check is assigned to the order type. However, there is one assigned for the delivery type. Also, for CCAr/Risk cat./CG combination in OVA8 (Auto Credit Block Delivery), static credit check has been activated with error message reaction. Open orders and open deliveries were also activated. Customer exceeded credit exposure but user cannot release the order in any VKMx transactions as no credit limit check is assigned to the order type.
    Given that this is the current setup, and the following conditions:
    1. Credit limit for the customer cannot be increased
    2. Oldest open item cannot be cleared just yet
    Is there any way that we can proceed with delivery creation for this customer? The other order types have credit limit check so credit management team was able to release them. However, we are unsure on how to proceed with this consignment order. Kindly advise and thank you in advance for your help.

    We are encountering a credit limit exceeded error in delivery creation for a consignment issue.
    but user cannot release the order in any VKMx transactions as no credit limit check is assigned to the order type.
    Looks contradictory and it may need to run report RVKRED77 to reorganization of open credit values. Try to run the report in test system and then run in production system. After running the report, the use must be able to release in VKM* transaction.
    Regards,

  • Error while running query "time limit exceeding"

    while running a query getting error "time limit exceeding".plz help.

    hi devi,
    use the following links
    queries taking long time to run
    Query taking too long
    with hopes
    Raja Singh

Maybe you are looking for

  • Cannot connect to other computers on network. HELP!

    Using a G-5 dual 2.0 with 10.4.10 and I am having issues with networking and connecting to other computers on my network. This is an intermittent problem. When I try to connect <go> <connect to server><browse> the computers on the network do not show

  • Need help in Multi mapping

    Hi experts, we have a scenario need to send an EDI text file to 2 different receivers. We are receving the EDI file using the JMS adapter as Sender and File/FTP adapter as a receiver. We are using Multi mapping (includes Java mapping and Graphical )

  • Can't get the "find" function to work in Bridge CS4?

    Hello there, I need a bit of help with the Find funtion in Brdige CS4. Basically it doesn't work. I've got a folder with multiple files and keywords and the search brings up "no results" regardless what I do. What could be the problem?? I'm running W

  • Using MainStage with multiple MIDI inputs (two keyboards!)

    Hi Guys, I use two keyboards for my live gigs, and am thinking about moving all my sounds over to MainStage. Most of my recording uses Native Instruments and Logic's standard AU instruments, and then I have to find similar sounds on my Korg and Kurzw

  • Why do we have different partners KU?

    HI Guys It might sound like a dumb question but I am trying to understand why do we have different partner profiles for KU type. I understand they are different customers and hence different but... My scenario is we are posting ORDERS (IDOCS) for 6 d