Application Builder takes long time and Popup messages during Build

Hi All,
My application is slightly bigger with over 300 VI's. In between the build process I receive pop up messages asking GMath library  files etc.. are modifies would u like to save the files..? When clicked on OK button it starts saving these library files and when it saves almost half the number of files the application builder crashes and causes LabView to close. As the build process is taking more than 2 hours and end result is always an unsuccesful build due to the application builder crash I am loosing my time.
Regards,
Pavan
Solved!
Go to Solution.

Hi,
I recently upgraded from LV 7.1 to 2009, both the Professional versions. My program consists of several hundred vi's. Building an .exe in LV 7.1 was a snap and took maybe 30-50 seconds (not including an installer). Building the same thing in LV 2009 (build script converted to project) takes nearly 30 minutes, not including creating an installer... most of the time is spent 'processing' and then saving vi's, which LV 7 did not appear to do. I've tried the Ctrl+Shift+Run suggested by JB but this does not help. I've also applied the 2009f patch. It still takes a fair amount of memory, 0.5 GB, but my core 2 duo PC has 2 GB and there is plenty available RAM. Any suggestions/details on the differences and something I could do to cut down on the build time?
thanks,
Dan

Similar Messages

  • Application Load Take Long time

    Hii all,
    i has developed and adf application and oracle 9i database and deployed on a weblogic 10.3.1 using JDBC Data Source an thin Oracle Driver for service connection and then when i run the application from the intranet in the other far building the application loading takes long time and then when loaded it works good so why this long time in loading and how i can fix it or how i can trace it and know where's the problem,, hope anyone can help me because the user wait for 5 minutes to load the application and he got bored my boss will kill me..
    Regards

    Ahmed,
    Have you done any detective work yourself to find out where the long load time is coming from? Are you testing this over a slow network link? If so, have a look at section A9 [url http://download.oracle.com/docs/cd/E15523_01/web.1111/b31973/ap_config.htm#ADFUI10059]of this part of the documentation to see how you can perhaps reduce the initial load time.
    If it's not the network, then you can give us some more information about where the long load times are.
    John

  • Problem Export to Excel it takes long time and Takes more space.

    Hello All,
    when we export to Excel in portal(reports), it takes long time and it takes more space.
    how to overcome this problem please any one knows provide the proper solution for this issues
    Regards,
    Ch.

    Hi Chetans,
    I have had the same problem, and I had to create a OSS message to SAP in order to solve this issue, until now I don't have an answer. They made me to change a lot of configuration in Java and BW without luck. To tell you more, when we try to export to excel the java instance restarts aumotatically.
    But we have this problem, when we try to export a huge quantity of data, so I found a note which says the limitation of exporting to excel. Take a special attention to the Performance section.
    Note 1178857
    [https://service.sap.com/sap/support/notes/1178857|https://service.sap.com/sap/support/notes/1178857]
    I recomend you that you create a message to SAP. If you find a solution, please let me know.
    Regards, Federico

  • HT1351 for syncing it takes long time and not completed what to do???

    for syncing it takes long time and not completed what to do???

    Debbie:
    deborahfromwindsor wrote:
    he advises restarting by inserting the OSX disc and pressing down the C button to reboot from there then selecting disk utility, hard disk and repair.... Does he mean me to hold down the C key on the alpha keyboard or the ctrl key?
    Should I just ask for my money back??? If it is a simple repair do I just literally push the disc in, push the power button and hold down the C button?
    That's where I would begin, too, with
    Repair Disk
    Insert Installer disk and Restart, holding down the "C" key until grey Apple appears.
    Go to Installer menu (Panther and earlier) or Utilities menu (Tiger) and launch Disk Utility.
    Select your HDD (manufacturer ID) in the left panel.
    Select First Aid in the Main panel.
    (Check S.M.A.R.TStatus of HDD at the bottom of right panel, and report if it saysanything but Verified)
    Click Repair Disk on the bottom right.
    If DU reports disk does not need repairs quit DU and restart.
    If DU reports errors Repair again and again until DU reports disk is repaired.
    If DU reports errors it cannot repair you will need touse autility like TechTool Pro,Drive Geniusor DiskWarrior
    First we need to determine if the issue you are experiencing with the computer is software or hardware based. Once we have gotten things sorted out there should be time enough to make you decision about keeping or returning it.
    cornelius

  • MDX report rendering takes long time and showing Conflict Message

    Hi All,
    This is my MDX Query
    with member
    [Measures].[Rent] as
    IIF(IsEmpty([Measures].[Budget]),
    NULL, [Measures].[Rent])
    select {[Measures].[Rent]}
    on columns,
                         [Property].[Address].[All].children *
             DESCENDANTS([Account].[Account Hierarchy].[Account Group].[Expenditures],
                         [Account].[Account Tree].[Account]) *
                         [Property].[Property].[All].children
    on rows
    from
       [Master]
    When i comment [Property] Dimension member , i am able to get the result, but i need Property Dimension in MDX.
    Can anyone give some idea ?
    Thanks in advance

    Hi Jarugulalaks,
    According to your description, it take long time to render the report when using [Property] dimension, right?
    In this case, the issue can be caused by that there are too many members under this dimension. In your query, you used CrossJoin function to join multiple dimensions which might cause the performance issue. If you cross-join medium-sized or large-sized sets
    (e.g., sets that contain more than 100 items each), you can end up with a result set that contains many thousands of items—enough to seriously impair performance. You can use the NonEmptyCrossjoin function instead of Corssjoin function. For the detail
    information, please see:
    http://sqlmag.com/data-access/cross-join-performance
    http://msdn.microsoft.com/en-us/library/ms144797.aspx
    Besides, the total time to generate a reporting server report can be divided into 3 elements: Time to retrieve the data (TimeDataRetrieval); Time to process the report (TimeProcessing); Time to render the report (TimeRendering). For the detail information,
    please refer to the link below to see Charlie's reply.
    http://social.msdn.microsoft.com/Forums/sqlserver/en-US/a0962b54-6fc2-4111-b8f1-a3722a65fa05/how-to-improve-performance-of-report?forum=sqlanalysisservices#a0962b54-6fc2-4111-b8f1-a3722a65fa05
    Regards,
    Charlie Liao
    TechNet Community Support

  • BEx query selection screen takes long time to popup

    Hi,
    I have a query which is taking long time to popup the selection screen. I have tried in RSRT the generate report option, the generating report takes 45 to 60 mins. After the report is generated the query selection screen comes up in couple of seconds. The first time I run the query everything its fine but then for the second time the problem comes back again.
    The selection screen takes +/- 40 mins to comes up in RSRT and in the portal it times out.
    I have run the statistics of the query and according to the statistics report the query takes a  lot of time for the following component :
    OLAP: Query Gen.     2681,312682
    I have removed the query from the cache monitor and after that regenerate the report. But these steps didnu2019t solve my problem.
    I made two test queries, in one query I have made the CKFu2019s and RKFu2019s directly at the infoprovider, and dragged and dropped in the KF structure. In the other query I made the CKFu2019s and RKFu2019s in the query KF structure itself. The query where I have the CKFu2019s and RKFu2019s at the infoprovider that query times out, the other query where I put the CKFu2019s and RKFu2019s in the query KF structure doesn't has this performance problem.
    What I donu2019t know is why the query is failing when I use RKFu2019s and CKFu2019s from the infoprovider, the other query donu2019t have this problem.
    Please advise how to proceed.
    Thanks in advance,
    Arash

    Hi,
    Please check the following notes which should be useful.
    1387593 - Performance optimization for query change/generation
    1416737 - Performance optimization for query change/generation(2)
    1396485 - Performance optimization for query check/save
    Then please execute the SE38 report ANALYZE_RSZ_TABLES (see notes
    1472915, 1506560) and repair/correct the different tables to optmize
    the query definition tables.
    Regards,
    Amit

  • Validate map in FDM takes long time and finally expires

    Hi,
    I have an issue with FDM Mapping.
    When i click the "validate" option in workflow, the FDM processing takes long time (more than an hour) and finally expires.
    But when i open a new web interface, i am seeing a gold fish on validate - which means validate mapping has been done.
    As we are in production environmeny, Can someone quickly clarify me, why this happen.
    Thanks,
    Siva

    Hello Kelly,
    Something you said concerns me: +"Because we automated data loads for the large file, this issue is extremely low on the priority list."+
    In this type of scenario we notice customers/partners implementing FDM as an 'ETL' utility. FDM is not an ETL utility and should not be treated/used as such. In the event that you are doing this, Support has no control of the application as it is a mis-use of what FDM is for.
    Files that are being pushed into FDM should always be broken down into the smallest components(locations) and used as a 'front end,' user interface.... not a server-side/IT application. If you meet this criteria of FDM, then there is not much Support can do.
    If you do not think you are misusing FDM, I would highly suggest you open/create a Support SR if you do not use FDM in that manor. Consultants/Partners are functional/design based.... and not necessarily trained as Support Engineers. Therefore they might not have the skills required to make such determinations.
    Thank you,
    Oracle Global Customer Support

  • Query Saving takes long time and giving error

    Hi Gurus,
    I am creating one query that have lot of calculations (CKF & RKF).
    When I am trying to save this query it is taking long time and it is giving error like RFC_ERROR_SYSTEM_FAILURE, Query Designer must be restarted, further work not possible.
    Please give me the solution for this.
    Thanks,
    RChowdary

    Hi Chowdary,
    Check the following note: 316470.
    https://websmp130.sap-ag.de/sap(bD1lbiZjPTAwMQ==)/bc/bsp/spn/sapnotes/index2.htm?numm=316470
    The note details are:
    Symptom
    There are no authorizations to change roles. Consequently, the system displays no roles when you save workbooks in the BEx Analyzer. In the BEx browser, you cannot move or change workbooks, documents, folders and so on.
    Other terms
    BW 2.0B, 2.0A, 20A, 20B, frontend, error 172, Business Explorer,
    RFC_ERROR_SYSTEM_FAILURE, NOT_AUTHORIZED, S_USER_TCD, RAISE_EXCEPTION,
    LPRGN_STRUCTUREU04, SAPLPRGN_STRUCTURE, PRGN_STRU_SAVE_NODES
    Reason and Prerequisites
    The authorizations below are not assigned to the user.
    Solution
    Assign authorization for roles
    To assign authorizations for a role, execute the following steps:
    1. Start Transaction Role maintenance (PFCG)
    2. Select a role
    3. Choose the "Change" switch
    4. Choose tab title "Authorizations"
    5. Choose the "Change authorization data" switch
    6. Choose "+ Manually" switch
    7. Enter "S_USER_AGR" as "Authorization object"
    8. Expand "Basis: Administration"/"Authorization: Role check""
    9. From "Activity" select "Create or generate" and others like "Display" or "Change"
    10. Under "Role Name", enter all roles that are supposed to be shown or changed. Enter "*" for all roles.
    11. You can re-enter authorization object "S_USER_AGR" for other activities.
    Assign authorization for transactions
    If a user is granted the authorization for changing a role, he/she should also be granted the authorization for all transactions contained in the role. Add these transaction codes to authorization object S_USER_TCD.
    1. Start the role maintenance transaction (PFCG).
    2. Select a role.
    3. Click on "Change".
    4. Choose the "Authorizations" tab.
    5. Click on "Change authorization data".
    6. Click on "+ manually".
    7. Specify "S_USER_TCD" as "Authorization object".
    8. Expand "Basis - Administration"/"Authorizations: Transactions in Roles".
    9. Under "Transaction", choose at least "RRMX" (for BW reports), "SAP_BW_TEMPLATE" (for BW Web Templates), "SAP_BW_QUERY" (for BW Queries and/or "SAP_BW_CRYSTAL" (for Crystal reports) or "*". Values with "SAP_BW_..." are not transactions, they are special node types (see transaction code NODE_TYPE_DEFINITION).
    Using the SAP System Trace (Transaction ST01), you can identify the transaction that causes NOT_AUTHORIZED.
    Prevent user assignment
    Having the authorization for changing roles, the user is not only able to change the menu but also to assign users. If you want to prevent the latter, the user must loose the authorization for Transactions User Maintenance (SU01) and Role maintenance (PFCG).
    Z1>Note
    Refer to Note 197601, which provides information on the different display of BEx Browser, BEx Analyzer and Easy Access menu.
    Please refer to Note 373979 about authorizations to save workbooks.
    Check in the transaction ST22 for more details on the Query designer failure or query log file.
    With Regards,
    Ravi Kanth.
    Edited by: Ravi kanth on Apr 9, 2009 6:02 PM

  • Installing patch 5217019 takes long time and does not end

    Hi,
    I try to install patch 5217109 to my TEST instance.While installing patch the 8 workers are selected default.7 of workers are in statu of completed but 1 worker is waiting for long times and does not end.
    What can be my problem?

    Start time for file is: Mon Aug 10 2009 11:45:01
    sqlplus -s APPS/***** @/oracle/TEST/testappl/ad/11.5.0/admin/sql/adpcpcmp.pls APPLSYS &pw_fnd APPS &pw_apps &systempwd 8 1 NONE FALSE
    Arguments are:
    AOL_schema = APPLSYS, AOL_password = *****,
    Schema_to_compile = APPS, Schema_to_compile_pw = *****,
    SYSTEM_password = *****, Total_workers = 8, Logical_worker_num = 1
    Object_type_to_not_compile = NONE
    Use_stored_dependencies = FALSE
    Connected.
    Checking for previously-invalid objects which are now valid...
    PL/SQL procedure successfully completed.
    Elapsed: 00:00:00.03
    Commit complete.
    Elapsed: 00:00:00.00
    Deleting any errors from previous run of this logical worker
    0 rows deleted.
    Elapsed: 00:00:00.00
    Commit complete.
    Elapsed: 00:00:00.00
    Compiling invalid objects...

  • Un-Registering takes long time and fails!!!

    I am trying to un register a schema that ia already registered in the XML DB. I am using JDeveloper to un register and it really takes a long time to do this. and eventually fails.
    what is going on? what is broken?
    XML DB is flaky and unreliable.
    right?

    First make sure that all connections that have used the XML Schema are disconnected. Schema deletion cannot start until all sessions that are using the schema have been closed as it needs to get an exclusive lock on the SGA entrires related to the XML Schema.
    If there are a large number of rows in the table(s) associated with the XML Schema truncate the table before dropping the XML Schema. If there are a large number of XDB repository resources assocaited with the table, truncate the table and then delete the resources with DBMS_XDB.DELETERESOURCE() mode 2 or 4 to ignore errors and avoid dangling ref issues.
    To monitor the progress of deleteSchema itself connect as sys and execute select count(*) from all_objects where owner = 'XXXX'. XXXX should be the name of the owner of the database schema that was the owner of the XML schema. See if the number of objects owned by that user is decreasing. If the number of object is decreasing have patience.

  • Table Import Takes long time and still running

    Hi All,
    MY DB Version: 10.2.o
    OS: Windows Server 2003
    I am trying to import on table which i have the export dump file which i take using expdp previously when i load that table on the same host
    by using below command:
    expdp scott/tiger@db10g tables=EMP,DEPT directory=TEST_DIR dumpfile=EMP_DEPT.dmp logfile=expdpEMP_DEPT.log
    after that i zip that dump and move it to external usb and now i need that table i copy that table and unzip that that dump
    Command i am using to do the import is :
    impdp scott/tiger@db10g tables=EMP,DEPT directory=TEST_DIR dumpfile=EMP_DEPT.dmp logfile=impdpEMP_DEPT.log
    But the query of import is still runing even not showing any amount of rows to be imported.
    i already make the tablespace in which the table was previosuly before dropping
    but when i check the sapce of tablespace that is also not consuming
    one error i got preiviously while performing this task is:
    Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Master table "CDR"."SYS_IMPORT_TABLE_03" successfully loaded/unloaded
    Starting "CDR"."SYS_IMPORT_TABLE_03":  cdr/********@tsiindia directory=TEST_DIR dumpfile=CAT_IN_DATA_042012.DMP tables=CAT_IN_DATA_042012 logfile=impdpCAT_IN_DATA_042012.log
    ORA-39065: unexpected master process exception in RECEIVE
    ORA-39078: unable to dequeue message for agent MCP from queue "KUPC$C_1_20120622102502"
    Job "CDR"."SYS_IMPORT_TABLE_03" stopped due to fatal error at 12:10
    i done some google on the same proble
    and i check streams_pool_size it will show zero and then i make it to 48M and after that
    SQL> show parameter streams_pool_size;
    NAME                                 TYPE        VALUE
    streams_pool_size                    big integer 48M
    But still it takes time
    Any help

    1) check the session :
    SQL> select username,sid,serial#,status,event,seconds_in_wait,wait_time,state,module from v$session where username='CDR'
      2  ;
    USERNAME                              SID    SERIAL# STATUS   EVENT                                                    SECONDS_IN_WAIT  WAIT_TIME STATE               MODULE
    CDR                                    73          1 ACTIVE   wait for unread message on broadcast channel                                   3          0 WAITING
    CDR                                    75          1 ACTIVE   wait for unread message on broadcast channel                                  10          0 WAITING
    CDR                                    77          1 ACTIVE   wait for unread message on broadcast channel                                  10          0 WAITING
    CDR                                    81        313 ACTIVE   wait for unread message on broadcast channel                                 530          0 WAITING             impdp.exe
    CDR                                    87         70 ACTIVE   enq: SS - contention                                                1581          0 WAITING             toad.exe
    CDR                                    90       1575 ACTIVE   wait for unread message on broadcast channel                                   3          0 WAITING
    CDR                                    92       1686 ACTIVE   enq: SS - contention                                                 619          0 WAITING
    CDR                                    99          5 ACTIVE   PX Deq Credit: send blkd                                               0          0 WAITING             TOAD 9.1.0.62
    CDR                                   103          3 ACTIVE   direct path read                                                       0          0 WAITING             TOAD 9.1.0.62
    CDR                                   105          6 ACTIVE   direct path read                                                       0          0 WAITING             TOAD 9.1.0.62
    CDR                                   107          6 ACTIVE   PX Deq Credit: send blkd                                               0          0 WAITING             TOAD 9.1.0.62
    USERNAME                              SID    SERIAL# STATUS   EVENT                                                    SECONDS_IN_WAIT  WAIT_TIME STATE               MODULE
    CDR                                   108         16 ACTIVE   PX Deq Credit: send blkd                                               1          0 WAITING             TOAD 9.1.0.62
    CDR                                   109         40 ACTIVE   PX Deq Credit: send blkd                                               1          0 WAITING             TOAD 9.1.0.62
    CDR                                   110          6 ACTIVE   enq: TX - row lock contention                                          1          0 WAITING             TOAD 9.1.0.62
    CDR                                   111         21 ACTIVE   direct path read                                                       0          0 WAITING             TOAD 9.1.0.62
    CDR                                   112         27 ACTIVE   PX Deq Credit: send blkd                                               1          0 WAITING             TOAD 9.1.0.62
    CDR                                   113          8 ACTIVE   log buffer space                                                       0          0 WAITING             TOAD 9.1.0.62
    CDR                                   117       4496 ACTIVE   db file scattered read                                                 0          0 WAITING             TOAD 9.1.0.62
    CDR                                   119          9 ACTIVE   PX Deq Credit: send blkd                                               0          0 WAITING             TOAD 9.1.0.62
    CDR                                   120         27 ACTIVE   PX Deq Credit: send blkd                                               0          0 WAITING             TOAD 9.1.0.62
    CDR                                   123          1 ACTIVE   sort segment request                                               22349          0 WAITING
    CDR                                   129         22 ACTIVE   PX Deq Credit: send blkd                                               0          0 WAITING             TOAD 9.1.0.62
    USERNAME                              SID    SERIAL# STATUS   EVENT                                                    SECONDS_IN_WAIT  WAIT_TIME STATE               MODULE
    CDR                                   131      14402 INACTIVE SQL*Net message from client                                         2580          0 WAITING             TOAD 9.1.0.62
    CDR                                   135         11 ACTIVE   log buffer space                                                       0          0 WAITING             TOAD 9.1.0.62
    CDR                                   136          6 ACTIVE   direct path read                                                       0          0 WAITING             TOAD 9.1.0.62
    CDR                                   138        234 ACTIVE   sort segment request                                               19859          0 WAITING
    CDR                                   162        782 INACTIVE SQL*Net message from client                                          550          0 WAITING             TOAD 9.1.0.62
    2) check the impprt status:
    SQL> select owner_name, job_name, operation, job_mode, state FROM dba_datapump_jobs;
    OWNER_NAME                     JOB_NAME                       OPERATION                      JOB_MODE                       STATE
    CDR                            SYS_IMPORT_TABLE_01            IMPORT                         TABLE                  EXECUTING
    3) in a new window
    C:\Documents and Settings\vikas>impdp cdr/cdr123_awcc@tsiindia dumpfile=CAT_IN_DATA_042012.dmp tables=CAT_IN_DATA_042012  logfile=impdpCAT_IN_DATA_042012.log directory=test_dir parallel=4
    Import: Release 10.1.0.2.0 - Production on Friday, 22 June, 2012 15:04
    Copyright (c) 2003, Oracle.  All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.1.0.2.0 - Production
    With the Partitioning, OLAP and Data Mining options
    Master table "CDR"."SYS_IMPORT_TABLE_01" successfully loaded/unloaded
    Starting "CDR"."SYS_IMPORT_TABLE_01":  cdr/********@tsiindia dumpfile=CAT_IN_DATA_042012.dmp tables=CAT_IN_DATA_042012 logfile=impdpCAT_IN_DATA_042012.log directory=test_dir parallel=4

  • Clear operation takes long time and gets interrupted in ThreadGate.doWait

    Hi,
    We are running Coherence 3.5.3 cluster with 16 storage enabled nodes and 24 storage disabled nodes. We have about hundred of partitioned caches with NearCaches (invalidation strategy = PRESENT, size limit for different caches 60-200K) and backup count = 1. For each cache we have a notion of cache A and cache B. Every day either A or B is active and is used by business logic while the other one is inactive, not used and empty. Daily we load fresh data to inactive caches, mark them as active (switch business logic to work with fresh data from those caches), and clear all yesterday's data in those caches which are not used today.
    So at the end of data load we execute NamedCache.clear() operation for each inactive cache from storage disabled node. From time to time, 1-2 times a week, the clear operation fails on one of 2 our biggest caches (one has 1.2M entries and another one has 350K entries). We did some investigations and found that NamedCache.clear operation fires many events within Coherence cluster to clear NearCaches so that operation is quite expensive. In some other simular posts there were suggestions to not use NamedCache.clear, but rather use NamedCache.destroy, however that doesn't work for us in current timelines. So we implemented simple retry logic that retries NamedCache.clear() operation up to 4 times with increasing delay between the attempts (1min, 2 min, 4 min).
    However that didn't help. 3 out of those attempts failed with the same error on one storage enabled node and 1 out of those 4 attempts failed on another storage enabled node. In all cases a Coherence worker thread that is executing ClearRequest on storage enabled node got interrupted by Guardian after it reached its timeout while it was waiting on lock object at ThreadGate.doWait. Please see below:
    Log from the node that calls NamedCache.clear()
    Portable(com.tangosol.util.WrapperException): (Wrapped: Failed request execution for ProductDistributedCache service on Member(Id=26, Timestamp=2012-09-04 13:37:43.922, Address=32.83.113.116:10000, MachineId=3149, Location=machine:mac305,process:2
    7091,member:mac305.instance1, Role=storage) (Wrapped: ThreadGate{State=GATE_CLOSING, ActiveCount=3, CloseCount=0, ClosingT
    hread= Thread[ProductDistributedCacheWorker:1,5,ProductDistributedCache]}) null) null
    Caused by:
    Portable(java.lang.InterruptedException) ( << comment: this came form storage enabled node >> )
    at java.lang.Object.wait(Native Method)
    at com.tangosol.util.ThreadGate.doWait(ThreadGate.java:489)
    at com.tangosol.util.ThreadGate.close(ThreadGate.java:239)
    at com.tangosol.util.SegmentedConcurrentMap.lock(SegmentedConcurrentMap.java:180)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache.onClearRequest(DistributedCache.CDB:27)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache$ClearRequest.run(DistributedCache.CDB:1)
    at com.tangosol.coherence.component.util.DaemonPool$WrapperTask.run(DaemonPool.CDB:1)
    at com.tangosol.coherence.component.util.DaemonPool$WrapperTask.run(DaemonPool.CDB:32)
    at com.tangosol.coherence.component.util.DaemonPool$Daemon.onNotify(DaemonPool.CDB:63)
    at com.tangosol.coherence.component.util.Daemon.run(Daemon.CDB:42)
    at java.lang.Thread.run(Thread.java:619)
    Log from the that storage enabled node which threw an exception
    Sat Sep 08 04:38:37 EDT 2012|**ERROR**| com.tangosol.coherence.component.util.logOutput.Log4j | 2012-09-08 04:38:37.720/31330
    1.617 Oracle Coherence EE 3.5.3/465 <Error> (thread=DistributedCache:ProductDistributedCache, member=26): Attempting recovery
    (due to soft timeout) of Guard{Daemon=ProductDistributedCacheWorker:1} |Client Details{sdpGrid:,ClientName:  ClientInstanceN
    ame: ,ClientThreadName:  }| Logger@9259509 3.5.3/465
    Sat Sep 08 04:38:37 EDT 2012|**WARN**| com.tangosol.coherence.component.util.logOutput.Log4j | 2012-09-08 04:38:37.720/313301
    .617 Oracle Coherence EE 3.5.3/465 <Warning> (thread=Recovery Thread, member=26): A worker thread has been executing task: Message "ClearRequest"
    FromMember=Member(Id=38, Timestamp=2012-09-07 10:12:27.402, Address=32.83.113.120:10000, MachineId=40810, Location=machine:
    mac313,process:22837,member:mac313.instance1, Role=maintenance)
    FromMessageId=5278229
    Internal=false
    MessagePartCount=1
    PendingCount=0
    MessageType=1
    ToPollId=0
    Poll=null
    Packets
    [000]=Directed{PacketType=0x0DDF00D5, ToId=26, FromId=38, Direction=Incoming, ReceivedMillis=04:36:49.718, ToMemberSet=nu
    ll, ServiceId=6, MessageType=1, FromMessageId=5278229, ToMessageId=337177, MessagePartCount=1, MessagePartIndex=0, NackInProg
    ress=false, ResendScheduled=none, Timeout=none, PendingResendSkips=0, DeliveryState=unsent, Body=0x000D551F0085B8DF9FAECE8001
    0101010204084080C001C1F80000000000000010000000000000000000000000000000000000000000000000, Body.length=57}
    Service=DistributedCache{Name=ProductDistributedCache, State=(SERVICE_STARTED), LocalStorage=enabled, PartitionCount=257, B
    ackupCount=1, AssignedPartitions=16, BackupPartitions=16}
    ToMemberSet=MemberSet(Size=1, BitSetCount=2
    Member(Id=26, Timestamp=2012-09-04 13:37:43.922, Address=32.83.113.116:10000, MachineId=3149, Location=machine:mac305,process:27091,member:mac305.instance1, Role=storage)
    NotifySent=false
    } for 108002ms and appears to be stuck; attempting to interrupt: ProductDistributedCacheWorker:1 |Client Details{sdpGrid:,C
    lientName: ClientInstanceName: ,ClientThreadName: }| Logger@9259509 3.5.3/465
    I am looking for your help. Please let me know if you see what is the reason for the issue and how to address it.
    Thank you

    Today we had that issue again and I have gathered some more information.
    Everything was the same as I described in the previous posts in this thread: first attempt to clear a cache failed and next 3 retries also failed. All 4 times 2 storage enabled nodes had that "... A worker thread has been executing task: Message "ClearRequest" ..." error message and got interrupted by Guardian.
    However after that I had some time to do further experiments. Our App has cache management UI that allows to clear any cache. So I started repeatedly taking thread dumps on those 2 storage enabled nodes which failed to clear the cache and executed cache clear operation form that UI. One of storage enabled nodes successfully cleared its part, but the other still failed. It failed with completely same error.
    So, I have a thread dump which I took while cache clear operation was in progress. It shows that a thread which is processing that ClearRequest is stuck waiting in ThreadGate.close method:
    at java.lang.Object.wait(Native Method)
    at com.tangosol.util.ThreadGate.doWait(ThreadGate.java:489)
    at com.tangosol.util.ThreadGate.close(ThreadGate.java:239)
    at com.tangosol.util.SegmentedConcurrentMap.lock(SegmentedConcurrentMap.java:180)
    at com.tangosol.coherence.component.util.daemon.queueProcessor.service.grid.DistributedCache.onClearRequest(DistributedCache.CDB:27)
    at
    All subsequents attempts to clear cache from cache management UI failed until we restarted that storage enabled node.
    It looks like some thread left ThreadGate in a locked state, and any further attempts to apply a lock as part of ClearRequest message fail. May be it is known issue of Coherence 3.5.3?
    Thanks

  • Starting Itunes match at atv2 takes long time and doesn't stop?

    I startetd itunes match at atv2. It says that it will take only a few minutes depending on the size of the mediathek. But it's running now for 2 days. What can I do?

    Do you mean iTunes Match has been running for several days on the initial match of your library?  If so, depending on the size of your library this is not unusual.
    If atv2 refers to Apple TV 2 then something is very wrong since this will start within minutes.

  • Report takes long time for few records

    hi frends,
    I m facing one problem with my Web based erp application which is developed in .net , in my application when i open the  report from my applicaiton , in my temp folder there one file gets created name is "rpt conmgr cache"
    bcoz of this for few records also my report takes too much time and opens very slow and it takes long time, and it happens in some of the reports only , other reports are working cool and its not creating any file in temp folder,,, so can u guide me whats this file and what can be the solution for it,
    Thanks
    Mithun

    hi sabhajit,
    i have already checked the sql query it is taking less then seconds.
    any other steps u want me to check then pls let me know?
    thanks mithun

  • Full GC takes long time

    Hello,
    Since the last two weeks, we're experiencing Full GCs that take long time and block our application.
    Our platform is:
    Windows 2003 Server 64 bit
    Tomcat 6.0.18
    Java 1.6.0_u13.
    Total memory: 8Gb.
    -Xms4096m
    -Xmx6144m
    -Xss256k
    Old generation: min 2,66Gb max 4,0G
    Eden generation: min 1,5Gb max 2,0G
    Perm generation: min 57Mb max 84Mb
    Here a piece of our gc.log:
    21601.266: [GC [PSYoungGen: 2009225K->6206K(1942528K)] 3913618K->1913365K(4738752K), 0.0992347 secs] [Times: user=0.16 sys=0.03, real=0.11 secs]
    21611.268: [GC [PSYoungGen: 1942485K->5485K(1879040K)] 3849645K->1914490K(4675264K), 0.0986129 secs] [Times: user=0.11 sys=0.00, real=0.11 secs]
    21618.350: [GC [PSYoungGen: 1879021K->12392K(2060928K)] 3788026K->1930794K(4857152K), 0.2036000 secs] [Times: user=0.14 sys=0.02, real=0.20 secs]
    21624.869: [GC [PSYoungGen: 1935116K->8545K(2065600K)] 3853519K->1930702K(4861824K), 0.0924182 secs] [Times: user=0.05 sys=0.00, real=0.09 secs]
    21624.961: [Full GC (System) [PSYoungGen: 8545K->0K(2065600K)] [PSOldGen: 1922156K->635750K(2796224K)] 1930702K->635750K(4861824K) [PSPermGen: 38643K->38643K(60992K)], 1.3719842 secs] [Times: user=1.09 sys=0.00, real=1.36 secs]
    21634.887: [GC [PSYoungGen: 2048500K->10395K(2078912K)] 2684250K->646145K(4875136K), 0.0180670 secs] [Times: user=0.02 sys=0.05, real=0.01 secs]
    21645.165: [GC [PSYoungGen: 2074007K->8120K(2003392K)] 2709758K->648930K(4799616K), 0.0176943 secs] [Times: user=0.08 sys=0.00, real=0.03 secs]
    25209.938: [GC [PSYoungGen: 1760190K->7344K(1707072K)] 4127764K->2376892K(4503296K), 0.0760053 secs] [Times: user=0.11 sys=0.03, real=0.06 secs]
    25218.120: [GC [PSYoungGen: 1707056K->10511K(1657728K)] 4076604K->2382964K(4453952K), 0.0704081 secs] [Times: user=0.13 sys=0.00, real=0.08 secs]
    25221.680: [GC [PSYoungGen: 1657679K->5731K(1982400K)] 4030132K->2385544K(4778624K), 0.0985384 secs] [Times: user=0.20 sys=0.00, real=0.09 secs]
    25226.968: [GC [PSYoungGen: 993426K->6083K(1982784K)] 3373239K->2389543K(4779008K), 0.5147901 secs] [Times: user=0.08 sys=0.00, real=0.52 secs]
    25227.483: [Full GC (System) [PSYoungGen: 6083K->0K(1982784K)] [PSOldGen: 2383460K->536034K(2796224K)] 2389543K->536034K(4779008K) [PSPermGen: 39765K->38756K(58816K)], 155.4030025 secs] [Times: user=1.34 sys=0.64, real=155.39 secs]
    25389.833: [GC [PSYoungGen: 1976624K->12915K(2023616K)] 2512658K->566671K(4819840K), 13.1263397 secs] [Times: user=0.11 sys=0.03, real=13.13 secs]
    25405.871: [GC [PSYoungGen: 2023579K->13857K(2031552K)] 2577335K->583736K(4827776K), 3.5796784 secs] [Times: user=0.03 sys=0.00, real=3.58 secs]
    25416.490: [GC [PSYoungGen: 2024544K->9059K(2074944K)] 2594423K->587581K(4871168K), 0.1184355 secs] [Times: user=0.06 sys=0.00, real=0.13 secs]
    25425.393: [GC [PSYoungGen: 2064144K->7415K(2075456K)] 2642665K->590222K(4871680K), 0.0441002 secs] [Times: user=0.02 sys=0.00, real=0.05 secs]
    25438.547: [GC [PSYoungGen: 2062519K->8857K(2075264K)] 2645326K->594735K(4871488K), 0.0301620 secs] [Times: user=0.01 sys=0.00, real=0.03 secs]
    The minor collections go perfect.
    The first full GC in this piece of log takes 1,36s, and the old generation passes from 1922156K->635750K
    But the second full GC (1 hour later) blocks the server for 155.39s, and the old generation pases from 2383460K->536034K
    The server load between execution of full GC was similar. Any differences I saw in this two moments:
    first full gc: number of Tomcat http threads : 230. The full gc provokes that Tomcat start new threads until get 500 (maxThreads)
    second full gc: number of Tomcat http threads : 500.
    I'm not an expert in tunning jvm. If somebody could help me, these pauses are killing the app and it's unacceptable for my client (for 2 min. the app does not work).
    If you need more information (threads stack, histograms, ... I have ...)
    Maybe I need to reduce the size of the old generation? Or change the garbage collector for old generation? The server cpu is going well, I prefer to consume more cpu in gc and avoid this pauses.
    Any help will be very very appreciated.
    Thanks in advance,
    Joan.

    Hello,
    Sorry for my continuous replies. Another full gc has occurred. This is the trace:
    262631.531: [GC [PSYoungGen: 2071685K->10889K(2078464K)] 4711353K->2652961K(4874688K), 0.3392037 secs] [Times: user=0.08 sys=0.00, real=0.34 secs]
    262640.608: [GC [PSYoungGen: 2071978K->11055K(2078784K)] 4714051K->2656136K(4875008K), 1.3267436 secs] [Times: user=0.06 sys=0.00, real=1.33 secs]
    262649.880: [GC [PSYoungGen: 2072165K->17337K(2071872K)] 4717246K->2669454K(4868096K), 3.0493937 secs] [Times: user=0.16 sys=0.00, real=3.05 secs]
    262658.069: [GC [PSYoungGen: 2071865K->14678K(2075840K)] 4723982K->2676078K(4872064K), 5.4598301 secs] [Times: user=0.08 sys=0.03, real=5.45 secs]
    262666.592: [GC [PSYoungGen: 2069183K->15835K(2075776K)] 4730583K->2682691K(4872000K), 1.6894894 secs] [Times: user=0.11 sys=0.05, real=1.70 secs]
    262673.352: [GC [PSYoungGen: 2070620K->14122K(2068992K)] 4737476K->2687230K(4865216K), 2.9711981 secs] [Times: user=0.06 sys=0.03, real=2.97 secs]
    262682.891: [GC [PSYoungGen: 2068909K->10709K(2075712K)] 4742017K->2686673K(4871936K), 1.1160323 secs] [Times: user=0.19 sys=0.00, real=1.13 secs]
    262691.733: [GC [PSYoungGen: 2065941K->10683K(2075584K)] 4741905K->2689212K(4871808K), 1.4898942 secs] [Times: user=0.05 sys=0.00, real=1.50 secs]
    262701.512: [GC [PSYoungGen: 2065888K->9341K(2076480K)] 4744418K->2690073K(4872704K), 1.3835846 secs] [Times: user=0.13 sys=0.00, real=1.39 secs]
    262709.896: [GC [PSYoungGen: 2065601K->10718K(2076480K)] 4746334K->2693212K(4872704K), 0.7815897 secs] [Times: user=0.05 sys=0.01, real=0.78 secs]
    262718.681: [GC [PSYoungGen: 2067028K->10199K(2076672K)] 4749523K->2695508K(4872896K), 1.2951014 secs] [Times: user=0.06 sys=0.00, real=1.30 secs]
    262727.940: [GC [PSYoungGen: 2067345K->10112K(2076608K)] 4752654K->2698129K(4872832K), 1.6785345 secs] [Times: user=0.08 sys=0.00, real=1.69 secs]
    262736.092: [GC [PSYoungGen: 2067264K->10886K(2077184K)] 4755281K->2702237K(4873408K), 1.1985764 secs] [Times: user=0.06 sys=0.02, real=1.20 secs]
    262744.626: [GC [PSYoungGen: 2069382K->11366K(2077248K)] 4760733K->2705795K(4873472K), 1.4323653 secs] [Times: user=0.13 sys=0.00, real=1.44 secs]
    262754.913: [GC [PSYoungGen: 2069912K->11945K(2077824K)] 4764342K->2708689K(4874048K), 0.8340946 secs] [Times: user=0.03 sys=0.00, real=0.84 secs]
    262765.038: [GC [PSYoungGen: 2072283K->11669K(2077888K)] 4769028K->2711204K(4874112K), 1.3040060 secs] [Times: user=0.09 sys=0.02, real=1.31 secs]
    262776.885: [GC [PSYoungGen: 2072021K->9940K(2079360K)] 4771556K->2711919K(4875584K), 0.8646772 secs] [Times: user=0.09 sys=0.00, real=0.88 secs]
    262783.713: [GC [PSYoungGen: 1173115K->8659K(2079168K)] 3875094K->2713127K(4875392K), 1.4434465 secs] [Times: user=0.09 sys=0.00, real=1.45 secs]
    262785.157: [Full GC (System) [PSYoungGen: 8659K->0K(2079168K)] [PSOldGen: 2704467K->655075K(2796224K)] 2713127K->655075K(4875392K) [PSPermGen: 41376K->40988K(42048K)], 132.6088125 secs] [Times: user=1.47 sys=0.39, real=132.58 secs]
    262925.100: [GC [PSYoungGen: 2062280K->16892K(2068992K)] 2717355K->683188K(4865216K), 8.9491611 secs] [Times: user=0.13 sys=0.00, real=8.94 secs]
    262971.183: [GC [PSYoungGen: 2068978K->8052K(2074624K)] 2735273K->692990K(4870848K), 7.7833800 secs]
    When a "bad" full GC occurs, I can see that just before it the minor collections start to perform bad. Normal values for minor collections are 0,0x seconds, and before a "bad" full GC these minor collections take seconds (1, 2, ... up to 5 seconds).
    I can show an histogram of the first 25 entries before and after the full GC:
    BEFORE
    num #instances #bytes class name
    1: 226135 718455272 [I
    2: 1905686 637239584 [C
    3: 122775 621626112 [B
    4: 8218 538714736 [Lcom.sun.org.apache.xpath.internal.objects.XObject;
    5: 160974 445403288 [Ljava.lang.Object;
    6: 1759928 70397120 java.lang.String
    7: 334672 42838016 com.sun.org.apache.xerces.internal.dom.DeferredElementNSImpl
    8: 103592 42699248 [[I
    9: 227966 41945744 java.net.SocksSocketImpl
    10: 629974 35278544 com.vpfw.proxy.cache.data.B
    11: 636515 30552720 java.util.concurrent.ConcurrentHashMap$HashEntry
    12: 555354 26656992 java.util.HashMap$Entry
    13: 89059 18034280 [Ljava.util.HashMap$Entry;
    14: 230432 16591104 com.sun.org.apache.xerces.internal.dom.DeferredTextImpl
    15: 225401 14425664 java.lang.ref.Finalizer
    16: 222588 14245632 java.net.SocketInputStream
    17: 334644 13385760 com.sun.org.apache.xerces.internal.dom.AttributeMap
    18: 386837 12378784 java.util.concurrent.ConcurrentLinkedQueue$Node
    19: 40226 11263504 [[Ljava.lang.Object;
    20: 21646 11246112 [Ljava.util.concurrent.ConcurrentHashMap$HashEntry;
    21: 684613 10953808 java.lang.Object
    22: 227958 9118320 java.net.Socket
    23: 218591 8743640 java.net.Inet4Address
    24: 57810 8730528 <constMethodKlass>
    25: 230658 7381056 java.io.FileDescriptor
    AFTER
    num #instances #bytes class name
    1: 91917 650028776 [B
    2: 1767088 612264368 [C
    3: 176313 581838552 [I
    4: 141278 322132688 [Ljava.lang.Object;
    5: 1762956 70518240 java.lang.String
    6: 367122 46991616 com.sun.org.apache.xerces.internal.dom.DeferredElementNSImpl
    7: 239506 44069104 java.net.SocksSocketImpl
    8: 633882 35497392 com.vpfw.proxy.cache.data.B
    9: 639797 30710256 java.util.concurrent.ConcurrentHashMap$HashEntry
    10: 606366 29105568 java.util.HashMap$Entry
    11: 72730 20521256 [[I
    12: 89136 18497696 [Ljava.util.HashMap$Entry;
    13: 252946 18212112 com.sun.org.apache.xerces.internal.dom.DeferredTextImpl
    14: 236339 15125696 java.lang.ref.Finalizer
    15: 233891 14969024 java.net.SocketInputStream
    16: 366906 14676240 com.sun.org.apache.xerces.internal.dom.AttributeMap
    17: 429913 13757216 java.util.concurrent.ConcurrentLinkedQueue$Node
    18: 43319 12129544 [[Ljava.lang.Object;
    19: 719233 11507728 java.lang.Object
    20: 21649 11246408 [Ljava.util.concurrent.ConcurrentHashMap$HashEntry;
    21: 239498 9579920 java.net.Socket
    22: 143 9317736 [Lcom.sun.org.apache.xpath.internal.objects.XObject;
    23: 229531 9181240 java.net.Inet4Address
    24: 57817 8735944 <constMethodKlass>
    25: 242443 7758176 java.io.FileDescriptor
    The big differencies is the "com.sun.org.apache.xpath.internal.objects.XObject;".
    Before and after, the values are:
    4: 8218 538714736 [Lcom.sun.org.apache.xpath.internal.objects.XObject;
    22: 143 9317736 [Lcom.sun.org.apache.xpath.internal.objects.XObject;
    Could this the cause of the full GC that takes long time?
    Thanks,
    Joan.

Maybe you are looking for