200 Point Limit Exceeded

We only have about 30 actual datapoints, but the PLC is handling all engineering unit conversion and alarm checking.  So for each datapoint requires that Lookout set or monitor 5 real values (Hi and Lo setpoints, Engr. conversion constants, and measured value) and 3 flags (hi and lo enable alarms and alarm indicator).  Which adds up to about 250 references on Modbus Plus which exceeds our 200 licenses limit.  How can I get Lookout to recognoze that there are only 30 actual datapoints??  We are using Lookout V5.1
If this is not possible, how do I increase the license point limit and how much does it cost???

Hello, In fact you do are using the 250 points that you mention. An IO point is not only the members for the PLC, but anything that has access to Citadel and Drivers.
Check the following Link to learn how many IO points you have. Said this, you need to upgrade your license for more IO points.
Ricardo S.
National Instruments

Similar Messages

  • PI 7.11 - SXI_CACHE Issue - SQL 0904 - Resource Limit Exceeded

    Hi PI & IBM I Gurus,
    We are having the SXI_CACHE issue in our PI 7.11 SPS04 system. When we try to do the delta cache refresh through SXI_CACHE, it is returning the error SQL 0904 - Resource limit exceeded. When we try to do the full cache refresh, it is getting the issue 'Application issue during request processing'.
    We have cleaned up the SQL Packages with DLTR3PKG command, which did not resolve the issue. We recently performed a system copy to build the QA instance and I observed that the adapter engine cache for the development is presented in the QA instance and removed that cache from there.
    I am not seeing the adapter engine connection data cache in our PI system. The adapter engine cache is working fine.
    All the caches are working fine from the PI Administration page. The cache connectivity test is failing with the same error as I mentioned for the SXI_CACHE.
    Please let me know if you have encountered any issue like this on IBM I 6.1 Platform.
    Your help is highly appreciated.
    Thanks
    Kalyan

    Hi Kalyan,
    SQL0904 has different reason codes ...
    Which one are you seeing ?
    Is the SQL pack really at its boundary of 1GB ?
    ... otherwise, it is perhaps a totally different issue ... then DLTR3PKG cannot help at all ...
    If you should see this big SQL Package, you should use PRTSQLINF in order to see if there is more or less over and over the same SQL in, just with different host variables or so ...
    If the last point should be the case, I would open a message with BC-DB-DB4 so that they can check how to help here or to talk to the application people to behave a bit different ...
    Regards
    Volker Gueldenpfennig, consolut international ag
    http://www.consolut.com http://www.4soi.de http://www.easymarketplace.de

  • PI 7.11 - SXI_CACHE Issue - SQL0904 - Resource limit exceeded

    Hi IBM I Gurus,
    We are having the SXI_CACHE issue in our PI 7.11 SPS04 system. When we try to do the delta cache refresh through SXI_CACHE, it is returning the error SQL 0904 - Resource limit exceeded. When we try to do the full cache refresh, it is getting the issue 'Application issue during request processing'.
    We have cleaned up the SQL Packages with DLTR3PKG command, which did not resolve the issue. We recently performed a system copy to build the QA instance and I observed that the adapter engine cache for the development is presented in the QA instance and removed that cache from there.
    I am not seeing the adapter engine connection data cache in our PI system. The adapter engine cache is working fine.
    All the caches are working fine from the PI Administration page. The cache connectivity test is failing with the same error as I mentioned for the SXI_CACHE.
    Please let me know if you have encountered any issue like this on IBM I 6.1 Platform.
    Your help is highly appreciated.
    Thanks
    Kalyan

    Hi Kalyan,
    SQL0904 has different reason codes ...
    Which one are you seeing ?
    Is the SQL pack really at its boundary of 1GB ?
    ... otherwise, it is perhaps a totally different issue ... then DLTR3PKG cannot help at all ...
    If you should see this big SQL Package, you should use PRTSQLINF in order to see if there is more or less over and over the same SQL in, just with different host variables or so ...
    If the last point should be the case, I would open a message with BC-DB-DB4 so that they can check how to help here or to talk to the application people to behave a bit different ...
    Regards
    Volker Gueldenpfennig, consolut international ag
    http://www.consolut.com http://www.4soi.de http://www.easymarketplace.de

  • Memory Leak, Receiver Got Null Message & Consumer limit exceeded on destina

    When running program that adds an Object message to a JMS queue and then recieves it. I get the following.
    1) interminitent NULL messages recieved.
    2) jms.JMSException: [C4073]: Consumer limit exceeded on destination interactionQueueDest. Even though only one receiver can be receiving via the supplied program.
    3) After many message are added to the queue 1000's the Message Queue goes to Out Of Memory exception. It should swap to disk!!
    STEPS TO FOLLOW TO REPRODUCE THE PROBLEM :
    RUN this program via a JSP call in the application server.
    JSP
    <%@ page language="java" import="jms.*"%>
    <html>
    <head>
    <title>Leak Memory</title>
    </head>
    <body>
    <hr/>
    <h1>Leak Memory</h1>
    <%
       LeakMemory leakMemory = new LeakMemory();
       leakMemory.runTest(10000,1000);
      // NOTE will brake but slower with setting leakMemory.runTest(10000,100);
    %>JMS resources must be created:
    jms/queueConnectionFactory
    jms/interactionQueue
    must be created first.
    Class:
    package jms;
    import javax.naming.*;
    import javax.jms.*;
    public class LeakMemory implements Runnable {
      private QueueConnectionFactory queueConnectionFactory = null;
      private Queue interactionQueue = null;
      private boolean receiverRun = true;
      private QueueConnection queueConnection;
      private int totalMessageInQueue = 0;
      public LeakMemory() {
        init();
       * initialize queues
      public void init(){
        try {
          InitialContext context = new InitialContext();
          this.queueConnectionFactory = (QueueConnectionFactory)context.lookup("jms/queueConnectionFactory");
          this.interactionQueue = (Queue) context.lookup("jms/interactionQueue");
        catch (NamingException ex) {
          printerError(ex);
      public void runTest(int messageCount, int messageSize){
        this.receiverRun = true;
        Thread receiverThread = new Thread(this);
        receiverThread.start();
        for (int i = 0; i < messageCount; i++) {
            StringBuffer messageToSend = new StringBuffer();
            for (int ii = 0; ii < messageSize; ii++) {
              messageToSend.append("0123456789012345678901234567890123456789012345678901234567890123456789012345678901234567890123456789\n");
            QueueConnection queueConnectionAdder = null;
            QueueSession queueInteractionSession = null;
            QueueSender interactionQueueSender = null;
            try {
                //Get a queue connection
                queueConnectionAdder = this.getQueueConnection();
                queueInteractionSession = queueConnectionAdder.createQueueSession(false, Session.AUTO_ACKNOWLEDGE);
                interactionQueueSender = queueInteractionSession.createSender(this.interactionQueue);
                ObjectMessage objectMessage = queueInteractionSession.createObjectMessage(messageToSend);
                objectMessage.setStringProperty("PROPERTY", "" + System.currentTimeMillis());
                //Send object
                interactionQueueSender.send(objectMessage,DeliveryMode.PERSISTENT,5,0);
                totalMessageInQueue++;
                //Close Resources
                interactionQueueSender.close();
                queueInteractionSession.close();
                interactionQueueSender = null;
                queueInteractionSession = null;
            } catch (JMSException ex) {
               printerError(ex);
       * run
      public void run()  {
        while(this.receiverRun){
                try {
                  QueueSession interactionQueueSession = this.getQueueConnection().createQueueSession(false, Session.CLIENT_ACKNOWLEDGE);
                  QueueReceiver queueReceiver = interactionQueueSession.createReceiver(this.interactionQueue);
                  ObjectMessage message = (ObjectMessage)queueReceiver.receive(100);
                  if(message != null){
                    StringBuffer messageRecived = (StringBuffer)message.getObject();
                    //Simulate Doing Something
                    synchronized (this) {
                      try {
                        Thread.sleep(400);
                      catch (InterruptedException ex1) {
                        //Can Safely be ignored
                    message.acknowledge();
                    totalMessageInQueue--;
                  } else {
                    printerError(new Exception("Receiver Got Null Message"));
                  queueReceiver.close();
                  interactionQueueSession.close();
                  queueReceiver = null;
                  interactionQueueSession = null;
                catch (JMSException ex) {
                  printerError(ex);
        * Get's the queue Connection and starts it
        * @return QueueConnection The queueConnection
       public synchronized QueueConnection getQueueConnection(){
           if (this.queueConnection == null) {
             try {
               this.queueConnection = this.queueConnectionFactory.createQueueConnection();
               this.queueConnection.start();
             catch (JMSException ex) {
               printerError(ex);
         return this.queueConnection;
       private void printerError(Exception ex){
         System.err.print("ERROR Exception totalMessageInQueue = " + this.totalMessageInQueue + "\n");
         ex.printStackTrace();
    }Is there something wrong with the way I'm working with JMS or is it just this unreliable in Sun App Server 7 Update 3 on windows?

    1) interminitent NULL messages recieved.Thanks that explains the behavior. It was wierd getting null messages when I know there is messages in the queue.
    2) jms.JMSException: [C4073]: Consumer limit exceeded
    on destination interactionQueueDest. Even though only
    one receiver can be receiving via the supplied
    program. No other instances, Only this program. Try it yourself!! It works everytime on Sun Application Server 7 update 2 & 3
    heres the broker dump at that error point
    [14/Apr/2004:12:51:47 BST] [B1065]: Accepting: [email protected]:3211->admin:3205. Count=1
    [14/Apr/2004:12:51:47 BST] [B1066]:   Closing: [email protected]:3211->admin:3205. Count=0
    [14/Apr/2004:12:52:20 BST] [B1065]: Accepting: [email protected]:3231->jms:3204. Count=1 [14/Apr/2004:12:53:31 BST] WARNING [B2009]: Creation of consumer  from connection [email protected]:3231 on destination interactionQueueDest failed:
    B4006: com.sun.messaging.jmq.jmsserver.util.BrokerException: [B4006]: Unable to attach to queue queue:single:interactionQueueDest: a primary queue is already active
    3) After many message are added to the queue 1000's
    the Message Queue goes to Out Of Memory exception. It
    should swap to disk!!The broker runs out of memory. Version in use
    Sun ONE Message Queue                   Copyright 2002
    Version:  3.0.1 SP2 (Build 4-a)              Sun Microsystems, Inc.
    Compile:  Fri 07/11/2003         All Rights ReservedOut of memory snippet
    [14/Apr/2004:13:08:28 BST] [B1089]: In low memory condition, Broker is attempting to free up resources
    [14/Apr/2004:13:08:28 BST] [B1088]: Entering Memory State [B0022]: YELLOW from previous state [B0021]: GREEN  - current memory is 118657K, 60% of total memory
    [14/Apr/2004:13:08:38 BST] WARNING [B2075]: Broker ran out of memory before the passed in VM maximum (-Xmx) 201326592 b,  lowering max to currently allocated memory (200431976 b ) and trying to recover [14/Apr/2004:13:08:38 BST] [B1089]: In low memory condition, Broker is attempting to free up resources
    [14/Apr/2004:13:08:38 BST] [B1088]: Entering Memory State [B0024]:  RED  from previous state [B0022]: YELLOW - current memory is 128796K, 99% of total memory [14/Apr/2004:13:08:38 BST] ERROR [B3008]: Message 2073-192.168.0.50(80:d:b6:c4:d6:73)-3319-1081944517772 exists in the store already [14/Apr/2004:13:08:38 BST] WARNING [B2011]: Storing of JMS message from IMQConn[AUTHENTICATED,[email protected]:3319,jms:3282] failed:
    com.sun.messaging.jmq.jmsserver.util.BrokerException: Message 2073-192.168.0.50(80:d:b6:c4:d6:73)-3319-1081944517772 exists in the store already
    [14/Apr/2004:13:08:38 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:38 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:39 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:40 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:41 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:42 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:43 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:44 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:44 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:44 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:45 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:45 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:46 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:46 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:47 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:48 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:49 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:50 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:51 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:52 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:52 BST] WARNING [B2076]: Broker is rejecting new producers, because it is extremely low on memory
    [14/Apr/2004:13:08:53 BST] ERROR [B3107]: Attempt to free memory failed, taking more drastic measures : java.lang.OutOfMemoryError
    [14/Apr/2004:13:08:53 BST] ERROR unable to deal w/ error: 1
    [14/Apr/2004:13:08:53 BST] ERROR TRYING TO CLOSE [14/Apr/2004:13:08:53 BST] ERROR DONE CLOSING
    [14/Apr/2004:13:08:53 BST] [B1066]:   Closing: [email protected]:3319->jms:3282. Count=0

  • Time Limit exceeded Error while updating huge number of records in MARC

    Hi experts,
    I have a interface requirement in which third party system will send a big file say.. 3 to 4MB file into SAP. in proxy we
    used BAPI BAPI_MATERIAL_SAVEDATA to save the material/plant data. Now, because of huge amount of data the SAP Queues are
    getting blocked and causing the time limit exceeded issues. As the BAPI can update single material at time, it will be called as many materials
    as we want to update.
    Below is the part of code in my proxy
    Call the BAPI update the safety stock Value.
        CALL FUNCTION 'BAPI_MATERIAL_SAVEDATA'
          EXPORTING
            headdata                    = gs_headdata
            CLIENTDATA                  =
            CLIENTDATAX                 =
           plantdata                   = gs_plantdata
           plantdatax                  = gs_plantdatax
           IMPORTING
              return              = ls_return.
        IF ls_return-type <> 'S'.
          CALL FUNCTION 'BAPI_TRANSACTION_ROLLBACK'.
          MOVE ls_return-message TO lv_message.
    Populate the error table and process next record.
          CALL METHOD me->populate_error
            EXPORTING
              message = lv_message.
          CONTINUE.
        ENDIF.
    Can any one please let me know what could be the best possible approach for this issue.
    Thanks in Advance,
    Jitender
    Hi experts,
    I have a interface requirement in which third party system will send a big file say.. 3 to 4MB file into SAP. in proxy we
    used BAPI BAPI_MATERIAL_SAVEDATA to save the material/plant data. Now, because of huge amount of data the SAP Queues are
    getting blocked and causing the time limit exceeded issues. As the BAPI can update single material at time, it will be called as many materials
    as we want to update.
    Below is the part of code in my proxy
    Call the BAPI update the safety stock Value.
        CALL FUNCTION 'BAPI_MATERIAL_SAVEDATA'
          EXPORTING
            headdata                    = gs_headdata
            CLIENTDATA                  =
            CLIENTDATAX                 =
           plantdata                   = gs_plantdata
           plantdatax                  = gs_plantdatax
           IMPORTING
              return              = ls_return.
        IF ls_return-type <> 'S'.
          CALL FUNCTION 'BAPI_TRANSACTION_ROLLBACK'.
          MOVE ls_return-message TO lv_message.
    Populate the error table and process next record.
          CALL METHOD me->populate_error
            EXPORTING
              message = lv_message.
          CONTINUE.
        ENDIF.
    Can any one please let me know what could be the best possible approach for this issue.
    Thanks in Advance,
    Jitender

    Hi Raju,
    Use the following routine to get fiscal year/period using calday.
    *Data definition:
    DATA: l_Arg1 TYPE RSFISCPER ,
          l_Arg2 TYPE RSFO_DATE ,
          l_Arg3 TYPE T009B-PERIV .
    *Calculation:
    l_Arg2  = TRAN_STRUCTURE-POST_DATE. (<b> This is the date that u have to give</b>)
    l_Arg3  = 'V3'.
    CALL METHOD CL_RSAR_FUNCTION=>DATE_FISCPER(
      EXPORTING I_DATE = l_Arg2
                I_PER = l_Arg3
      IMPORTING E_FISCPER = l_Arg1  ).
    RESULT = l_Arg1 .
    Hope it will sove ur problem....!
    Please Assign points.......
    Best Regards,
    SG

  • Program terminated: Time limit exceeded, ABAP performance, max_wprun_time

    Hi,
    I am running an ABAP program, and I get the following short dump:
    Time limit exceeded. The program has exceeded the maximum permitted runtime and has therefore been terminated. After a certain time, the program terminates to free the work processfor other users who are waiting. This is to stop work processes being blocked for too long by
    - Endless loops (DO, WHILE, ...),
    - Database acceses with large result sets,
    - Database accesses without an apporpriate index (full table scan)
    - database accesses producing an excessively large result set,
    The maximum runtime of a program is set by the profile parameter "rdisp/max_wprun_time". The current setting is 10000 seconds. After this, the system gives the program a second chance. During the first half (>= 10000 seconds), a call that is blocking the work process (such as a long-running SQLstatement) can occur. While the statement is being processed, the database layer will not allow it to be interrupted. However, to stop the program terminating immediately after the statement has been successfully processed, the system gives it another 10000 seconds. Hence the maximum runtime of a program is at least twice the value of the system profile parameter "rdisp/max_wprun_time".
    Last error logged in SAP kernel
    Component............ "NI (network interface)"
    Place................ "SAP-Dispatcher ok1a11cs_P06_00 on host ok1a11e0"
    Version.............. 34
    Error code........... "-6"
    Error text........... "connection to partner broken"
    Description.......... "NiPRead"
    System call.......... "recv"
    Module............... "niuxi.c"
    Line................. 1186
    Long-running programs should be started as background jobs. If this is not possible, you can increase the value of the system profile parameter "rdisp/max_wprun_time".
    Program cannot be started as a background job. We have now identified two options to solve the problem:
    - Increase the value of the system profile parameter "rdisp/max_wprun_time"
    - Improve the performance of the following SELECT statement in the program:
    SELECT ps_psp_pnr ebeln ebelp zekkn sakto FROM ekkn
    INTO CORRESPONDING FIELDS OF TABLE i_ekkn
    FOR ALL ENTRIES IN p_lt_proj
    WHERE ps_psp_pnr = p_lt_proj-pspnr
    AND ps_psp_pnr > 0.
    In EKKN we have 200 000 entries.
    Is there any other options we could try?
    Regards,
    Jarmo

    Thanks for your help, this problem seems to be quite challenging...
    In EKKN we have 200 000 entries. 199 999 entries have value of 00000000 in column ps_psp_pnr, and only one has a value which identifies a WBS element.
    I believe the problem is that there isn't any WBS element in PRPS which has the value of 00000000. I guess that is the reason why EKKN is read sequantially.
    I also tried this one, but it doesn't help at all. Before the SELECT statement is executed, there are 594 entries in internal table p_lt_proj_sel:
      DATA p_lt_proj_sel LIKE p_lt_proj OCCURS 0 WITH HEADER LINE.
      p_lt_proj_sel[] = p_lt_proj[].
      DELETE p_lt_proj_sel WHERE pspnr = 0.
      SORT p_lt_proj_sel by pspnr.
      SELECT ps_psp_pnr ebeln ebelp zekkn sakto FROM ekkn
      INTO CORRESPONDING FIELDS OF TABLE i_ekkn
      FOR ALL ENTRIES IN p_lt_proj_sel
      WHERE ps_psp_pnr = p_lt_proj_sel-pspnr.
    I also checked that the index P in EKKN is active.
    Can I somehow force the optimizer to use the index?
    Regards,
    Jarmo

  • TRFC error "time limit exceeded"

    Hi Prashant,
    No reply to my below thread...
    Hi Prashant,
    We are facing this issue quite often as i stated in my previous threads.
    As you mentioned some steps i have already followed all the steps so that i can furnish the jog log and tRFC details for reference long back.
    This issue i have posted one month back with full details and what we temporarily follow to execute this element successfully.
    Number of times i have stated that i need to know the root cause and permanent solution to resolve this issue as the log clearly states that it is due to struck LUWs(Source system).
    Even after executing the LUWs manually the status is same (Request still running and the status is in yellow color).
    I have no idea why this is happening to this element particularly as we have sufficient background jobs.
    we need change some settings like increasing or decreasing data package size or something else to resolve the issue permanently?
    For u i am giving the details once again
    Data flow:Standard DS-->PSA--->Data Target(DSO)
    In process monitor screen the request is in yellow color. NO clear error message s defined here.under update 0 record updated and missing message with yellow color except this the status against each log is green.
    Job log:Job is finished TRFCSSTATE=SYSFAIL message
    Trfcs:time limit exceeded
    What i follow to resolve the issue:Make the request green and manually update from PSA to Data target and the job gets completed successfully.
    Can you please tell me how to follow in this scenario to resolve the issue as i waiting for the same for long time now.
    And till now i didn't get any clue and what ever i have investigated i am getting replies till that point and no further update beyond this
    with regards,
    musai

    Hi,
    You have mentioned that already you have checked for LUWs, so the problem is not there now.
    In source system, go to we02 and check for idoc of type RSRQST & RSINFO. If any of them are in yellow status, take them to BD87 and process them. If the idoc processed is of RSRQST type, it would now create the job in source system for carrying out dataload. If it was of RSINFO type, it would finish the dataload in SAP BI side as well.
    If any in red, then check the reason.

  • Error "time limit Exceed"?"

    hi Experts,
    What to do when a load is failing with "time limit Exceed"?

    Hi,
    Time outs can be due to many reasons.  You will need to find out for your specific build.  Some of the common ones are:
    1. You could have set a large packet size.  See in the monitor whether the number of records in one packet seem inordinately large, say 100,000.  Reduce the number in steps and see which ones work for you.
    2.  The target may have a large number of fields, even then you will receive a time out as the size of the packet may become large.  Same solution as point 1.
    3.  You may have built an index in the target ODS which may impact your write speed.  Remove any indexes and run with the same package size, if it works then you know the index is the problem.
    4.  There is a basis setting for time out.  Check that is set as per SAP recommendation for your system.
    5.  Check the transactional RFCs in the source system.  It may have choked due to a large number of errors or hung queues.
    Cheers...

  • Database error text.. "Resource limit exceeded. MSGID=  Job=753531/IBPADM/WP05"

    Hello All,
    We are getting below runtime errors
    Runtime Errors         DBIF_RSQL_SQL_ERROR
    Exception              CX_SY_OPEN_SQL_DB
    Short text
        SQL error in the database when accessing a table.
    What can you do?
        Note which actions and input led to the error.
        For further help in handling the problem, contact your SAP administrator
        You can use the ABAP dump analysis transaction ST22 to view and manage
        termination messages, in particular for long term reference.
    How to correct the error
        Database error text........: "Resource limit exceeded. MSGID=  Job=753531/IBPADM/WP05""
       Internal call code.........: "[RSQL/OPEN/PRPS ]"
        Please check the entries in the system log (Transaction SM21).
        If the error occures in a non-modified SAP program, you may be able to
        find an interim solution in an SAP Note.
        If you have access to SAP Notes, carry out a search with the following
        keywords:
        "DBIF_RSQL_SQL_ERROR" "CX_SY_OPEN_SQL_DB"
        "SAPLCATL2" or "LCATL2U17"
        "CATS_SELECT_PRPS"
        If you cannot solve the problem yourself and want to send an error
        notification to SAP, include the following information:
        1. The description of the current problem (short dump)
           To save the description, choose "System->List->Save->Local File
        (Unconverted)".
        2. Corresponding system log
           Display the system log by calling transaction SM21.
           Restrict the time interval to 10 minutes before and five minutes
    after the short dump. Then choose "System->List->Save->Local File
    (Unconverted)".
    3. If the problem occurs in a problem of your own or a modified SAP
    program: The source code of the program
        In the editor, choose "Utilities->More
    Utilities->Upload/Download->Download".
    4. Details about the conditions under which the error occurred or which
    actions and input led to the error.
    System environment
        SAP-Release 700
        Application server... "SAPIBP0"
        Network address...... "3.14.226.140"
        Operating system..... "OS400"
        Release.............. "7.1"
           Character length.... 16 Bits
        Pointer length....... 64 Bits
        Work process number.. 5
        Shortdump setting.... "full"
        Database server... "SAPIBP0"
        Database type..... "DB400"
        Database name..... "IBP"
        Database user ID.. "R3IBPDATA"
        Terminal................. "KRSNBRB032"
        Char.set.... "C"
        SAP kernel....... 721
        created (date)... "May 15 2013 01:29:20"
        create on........ "AIX 1 6 00CFADC14C00 (IBM i with OS400)"
        Database version. "DB4_71"
    Patch level. 118
    Patch text.. " "
    Database............. "V7R1"
    SAP database version. 721
    Operating system..... "OS400 1 7"
    Memory consumption
    Roll.... 0
    EM...... 12569376
    Heap.... 0
    Page.... 2351104
    MM Used. 5210400
    MM Free. 3166288
    Information on where terminated
        Termination occurred in the ABAP program "SAPLCATL2" - in "CATS_SELECT_PRPS".
        The main program was "CATSSHOW ".
        In the source code you have the termination point in line 67
        of the (Include) program "LCATL2U17".
        The termination is caused because exception "CX_SY_OPEN_SQL_DB" occurred in
        procedure "CATS_SELECT_PRPS" "(FUNCTION)", but it was neither handled locally
         nor declared
        in the RAISING clause of its signature.
        The procedure is in program "SAPLCATL2 "; its source code begins in line
        1 of the (Include program "LCATL2U17 ".
    The exception must either be prevented, caught within proedure
    "CATS_SELECT_PRPS" "(FUNCTION)", or its possible occurrence must be declared in
      the
    RAISING clause of the procedure.
    To prevent the exception, note the following:
    SM21: Log
    Database error -904 at PRE access to table PRPS
    > Resource limit exceeded. MSGID= Job=871896/VGPADM/WP05
    Run-time error "DBIF_RSQL_SQL_ERROR" occurred
    Please help
    Regards,
    Usha

    Hi Usha
    Could you check this SAP Notes
      1930962 - IBM i: Size restriction of database tables
    1966949 - IBM i: Runtime error DBIF_DSQL2_SQL_ERROR in RSDB4UPD
    BR
    SS

  • 200 message limit during one 24 hour period

    on several occasions I have been locked out of my ability to send messages, and have gotten a message saying that I had sent more than 200 messages in the last 24 hours. How can I change this so that I am not locked once I have sent that many messages-in this case I couldn't help it-I was required by my employer to forward him all of the messages I had sent in the last several months, and that added up to more than 200-is there a setting like Storage preferences which I can change so that I am not limited by this 200 message limit, which I have hit several times, for the same reason?
    IMAC   Mac OS X (10.4.9)   2 GB memory

    You're welcome for the answer anyway.
    Does your employer provide you an email account and if so, why aren't you using your employer's email account with Mail to do their business or for their request?
    ALL ISPs and email account providers have sending limits for non-business personal accounts - repeat ALL.
    My ISP does not have a total recipient limit for each message sent but does have an overall recipient limit for all messages sent in a 24 hour period which can be reached with a single message.
    There is also a limit for the number of messages sent in a 24 hour period but I've never reached or exceeded any such limits imposed by my ISP or by Apple with my .Mac account but I don't use my personal .Mac account for my employer's business or their requests.
    I'm certainly no baby and I've always used either account for personal use only which is what non-business email accounts are designed and intended for and the reason for such restrictions along with part of an overall effort to prevent bulk spam mailings emanating from an ISP's or email account provider's domain.
    Your employer should provide you an email account which should be a BUSINESS account and have no sending restrictions or more liberal restrictions.

  • Client license limit exceeded?

    Computer #1:
    Software: 500 I/O points
    Client License: 1
    Lookout reports:
    2 Processes Running (136 total I/O points)
    Process #1 (136 total points, 136 input points, 8 output
    points)
    Process #2 (0 total points, 0 input points, 0 output points)
    Computer #2:
    Software: 500 I/O points
    Client License: 1
    Lookout reports:
    2 Processes Running (136 total I/O points)
    Process #1 (136 total points, 136 input points, 8 output
    points)
    Process #2 (0 total points, 0 input points, 0 output points)
    Each computer is both a server and client to one another (for mutual
    redundancy).
    I keep getting the "Client License
    Limit Exceeded" alarm on Computer #1.
    Even when Computer #1 takes over all of Computer #2 I/O points - it is still
    only 272 I/O points. I have two client licenses.
    Why would I get this alarm?

    I am also receiving this error, what I am trying to do is this;
    Using two computers, both with development and runtime licenses and both running the same process gathering data from a remote PLC.  There is one occasion that I need to press a button on one process (which alters the state of a flip/flop object) and have it change the state of the flip/flop on the other process. example parameters;
    flip/flop name:  S14_Trouble_ff
    input: S14_Trouble_pb or \\192.168.100.100\process_name\S14_Trouble_pb
    When I create this connection to the remote/network process, I receive the Client License Limit Exceeded alarm.  Is there a way around this?
    Thanks

  • TRFC Time limit exceeded

    Hi guys,
    Data loading takes place from ODS to cube is stuck up every day because of Time limit exceeded in TRFC queue.
    it is full load and it contains about  370000 records everyday.
    I set data packet size 10,000 (48000 records per packet) , however it was throughing an error.
    usually it is tacking 3hrs to load from ODS to cube,
    1. why such long time its taking to load from ODS to cube.
    2. is there any memory issue
    3.any other suggestion to avoid data packet stuckup
    4.how to  reduce the loading time.
    or
    whether I can reduce data packet size less than 10,000 to avoid this stuckup or to reduce dtataload time.
    Many thanks
    Ram

    Hi Ram,
    We can increase the Background Process Jobs, it is the BASIS work. Please check it with the basis people they will do it for you..
    Processes are depends on the calculation of the RAM u have minimum is 2 and the maximum is 18 work process per an instance. For the same RAM should have calculated for update process with minimum of 1 and the maximum is 3 background processes with 1-6... Calculations comes like.....RAM/256 for work process and differ for other processes. If you have UNIX environment then you can use as many Processes as you want. Actually, 1 WP allocates approx. 15-20MB RAM when system is  idle.
    Hence forth see your RAM and analyze the Predecessors, then workout this with your BASIS people.
    Pls assign points if useful.
    Cheers!
    Ragahvendra Rao.Kolli

  • License.Limit.Exceeded with 4x150 users licence and only 300 real users

    We have 4x150 users licence, set to unlimited bandwith.
    We get these messages in event viewer:
    Connection rejected by server. Reason : [
    License.Limit.Exceeded ] : (_defaultRoot_, _defaultVHost_) : Max
    connections allowed exceeds license limit. Rejecting connection to
    : our_app_name/.
    It appends a lot since the last few days, even before last
    update 2.0.3
    At the same time, when we check the administration console,
    the number in the _defaultVhost is much lowe than 600, the maximum
    number of user that is allowed.
    Licences are all OK in the administration console.
    We have to restart the Flash server service to correct the
    problem.
    Please help...

    Hello mm_Patrick,
    Thank you for your interest in this problem.
    We restrict multiple access to the same application instance
    on our service.
    We operate a chat using fms. We experience, from time to
    time, an application instance failure which results in the entire
    room participants being disconnected. The application instance
    however still shows the connections to the room at time of this
    failure. The problem starts to compound itself as people try to
    access the room ( connect to the application instance ). After the
    start of this problem when we look in the admin console the live
    log for the application instance does not show any new connection
    attempts nor accept or failure to connect but the application
    instance registers the connection against the total connect to the
    application instance. People are unable to connect to this room. As
    they continue to try they increase the connection number registered
    for this application instance until the point that the application
    instance total connected takes us past our license limit. The
    application instance seems to register and collect the connections
    even though there is no connection to the application instance. We
    then start to see the logging of the license limit exceeded.
    Please note that there are no error/warning messages
    generated in logs that indicate anything has failed. We can only
    determine that a failure once we start to experience the symptoms
    described for the application instance.
    We have experienced this issue in the past with fcs. It was
    not until the last update of fcs that the probem went away.
    However, fms 2.0.0 onward has always had this issue for us.
    The problem with this issue is that it is difficult to create
    and study. We can go 1 hour and crash or we may have success for
    over two weeks.
    In an attempt to search for a solution to the problem we have
    started focussing on this comment in Adobe –TechNote:
    http://www.adobe.com/go/d47f06c6
    “In Server.xml, find the ResourceLimits node (around
    line 181). It contains the Connector node (around line 203). Within
    the HTTP node (around line 205), create a new line and insert
    MaxConnectionThreads and give it an appropriate value:
    <MaxConnectionThreads>20</MaxConnectionThreads>
    The default is 10.”
    We have found that by adding the change to our HTTP node that
    we have lessened the frequency of our problem. Also, we have taken
    all remoting away from the time of connection. We now accept all
    connections and then do our work after connection.
    Another step we have taken to minimize the impact of this
    problem is to define the scope as “inst” in the
    application.xml. By creating a separate fmscore.exe for each
    application instance it gives us the ability to correct a problem
    with a faulting application instances without disturbing our entire
    server traffic.
    Any further insight into this problem would be greatly
    appreciated.
    Regards,
    iGx

  • Sending limit exceeded

    Hi,
    Every couple of days I seem to be getting the "daily sending limit exceeded" message when I try to send mail.  I'm aware that the limit is 200 messages per day, but I haven't send even close to that.  In fact, I've probably sent about 200 messages this month!
    Really can't understand why, but it's pretty frustrating.
    Edit - should note that it's my iCloud email that's giving me this error.  *@me.com account.

    Hey greigb99!
    I have an article for you that has some information regarding this message and how to proceed:
    iCloud: Mailbox size and message sending limits
    http://support.apple.com/kb/ht4863
    If you continue to exceed message sending limits, try the following methods to resolve the issue:
    Check your Mail Outbox to see if it contains a backlog of messages. Resend or delete messages here that may be continuing to try to send.
    Check to see if you have any network monitoring software that might be using your iCloud account to automatically send messages, such as firewalls or internet security software configured to send an alert when a potential security issue is detected.
    If you send messages to groups, make sure all email addresses in the group are valid. (For example, is the recipient's email address spelled correctly?
    Thanks for using the Apple Support Communities. Have a good one!
    -Braden

  • RFC_ERROR_SYSTEM_FAILURE: Time limit exceeded. Connection Pool - JCO api

    Hi Everyone
    My Connection  Pool parameters JCO api.
    client=300
    user=SISGERAL_RFC
    passwd=******
    ashost=14.29.3.120
    sysnr=00
    size=10
    I have these parameters on my Connection Pool and sometimes appear these wrongs in my application:
    1.
    2006-01-07 13:20:37,414 ERROR com.tel.webapp.framework.SAPDataSource - ##### Time limit exceeded. LOCALIZED MESSAGE = Time limit exceeded. KEY = RFC_ERROR_SYSTEM_FAILURE GROUP = 104 TOSTRING = com.sap.mw.jco.JCO$Exception: (104) RFC_ERROR_SYSTEM_FAILURE: Time limit exceeded.
    2.
    2006-01-07 14:01:31,007 ERROR com.tel.webapp.framework.SapPoolConnectionManager - Timeout
    I’d like to know if is happening.
    Are there something wrong with my connection pool?
    What can be happening?
    Thanks

    Raghu,
    Thanks for your response.
    Yes, the pool connections are in place according to the sAP note mentioned above.
    Regards,
    Faisal

Maybe you are looking for