Update statement time too long

HI
im doing this update statement using toad and its been an hour now and not finished ( its only 900 records)
and when i try to update 20 records only it takes about 3min
update IC_ITEM_MST set WHSE_ITEM_ID ='503' where ITEM_NO like 'PP%'
thnx
Edited by: george samaan on Dec 21, 2008 10:35 PM

select * from v$locked_object
gave me this
XIDUSN XIDSLOT XIDSQN OBJECT_ID SESSION_ID ORACLE_USERNAME
OS_USER_NAME PROCESS LOCKED_MODE
11 15 10999 36834 97 APPS
appltst2 897256 3
10 47 347465 63200 14 APPS
Administrator 3124:2324 2
10 47 347465 63569 14 APPS
Administrator 3124:2324 2
10 47 347465 63867 14 APPS
Administrator 3124:2324 3
10 47 347465 64380 14 APPS
Administrator 3124:2324 2
10 47 347465 64447 14 APPS
Administrator 3124:2324 2
10 47 347465 64934 14 APPS
XIDUSN XIDSLOT XIDSQN OBJECT_ID SESSION_ID ORACLE_USERNAME
OS_USER_NAME PROCESS LOCKED_MODE
Administrator 3124:2324 2
10 47 347465 78678 14 APPS
Administrator 3124:2324 3
10 47 347465 79069 14 APPS
Administrator 3124:2324 3
10 47 347465 64026 14 APPS
Administrator 3124:2324 3
10 47 347465 93468 14 APPS
Administrator 3124:2324 3
10 47 347465 209903 14 APPS
Administrator 3124:2324 3
10 47 347465 80084 14 APPS
Administrator 3124:2324 3
XIDUSN XIDSLOT XIDSQN OBJECT_ID SESSION_ID ORACLE_USERNAME
OS_USER_NAME PROCESS LOCKED_MODE
0 0 0 36944 60 APPS
appltst2 1572894 3
14 rows selected.

Similar Messages

  • SQL Update statement taking too long..

    Hi All,
    I have a simple update statement that goes through a table of 95000 rows that is taking too long to update; here are the details:
    Oracle Version: 11.2.0.1 64bit
    OS: Windows 2008 64bit
    desc temp_person;
    Name                                                                                Null?    Type
    PERSON_ID                                                                           NOT NULL NUMBER(10)
    DISTRICT_ID                                                                     NOT NULL NUMBER(10)
    FIRST_NAME                                                                                   VARCHAR2(60)
    MIDDLE_NAME                                                                                  VARCHAR2(60)
    LAST_NAME                                                                                    VARCHAR2(60)
    BIRTH_DATE                                                                                   DATE
    SIN                                                                                          VARCHAR2(11)
    PARTY_ID                                                                                     NUMBER(10)
    ACTIVE_STATUS                                                                       NOT NULL VARCHAR2(1)
    TAXABLE_FLAG                                                                                 VARCHAR2(1)
    CPP_EXEMPT                                                                                   VARCHAR2(1)
    EVENT_ID                                                                            NOT NULL NUMBER(10)
    USER_INFO_ID                                                                                 NUMBER(10)
    TIMESTAMP                                                                           NOT NULL DATE
    CREATE INDEX tmp_rs_PERSON_ED ON temp_person (PERSON_ID,DISTRICT_ID) TABLESPACE D_INDEX;
    Index created.
    ANALYZE INDEX tmp_PERSON_ED COMPUTE STATISTICS;
    Index analyzed.
    explain plan for update temp_person
      2  set first_name = (select trim(f_name)
      3                    from ext_names_csv
      4                               where temp_person.PERSON_ID=ext_names_csv.p_id
      5                               and   temp_person.DISTRICT_ID=ext_names_csv.ed_id);
    Explained.
    @?/rdbms/admin/utlxpls.sql
    PLAN_TABLE_OUTPUT
    Plan hash value: 3786226716
    | Id  | Operation                   | Name           | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | UPDATE STATEMENT            |                | 82095 |  4649K|  2052K  (4)| 06:50:31 |
    |   1 |  UPDATE                     | TEMP_PERSON    |       |       |            |          |
    |   2 |   TABLE ACCESS FULL         | TEMP_PERSON    | 82095 |  4649K|   191   (1)| 00:00:03 |
    |*  3 |   EXTERNAL TABLE ACCESS FULL| EXT_NAMES_CSV  |     1 |   178 |    24   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       3 - filter("EXT_NAMES_CSV"."P_ID"=:B1 AND "EXT_NAMES_CSV"."ED_ID"=:B2)
    Note
       - dynamic sampling used for this statement (level=2)
    19 rows selected.By the looks of it the update is going to take 6 hrs!!!
    ext_names_csv is an external table that have the same number of rows as the PERSON table.
    ROHO@rohof> desc ext_names_csv
    Name                                                                                Null?    Type
    P_ID                                                                                         NUMBER
    ED_ID                                                                                        NUMBER
    F_NAME                                                                                       VARCHAR2(300)
    L_NAME                                                                                       VARCHAR2(300)Anyone can help diagnose this please.
    Thanks
    Edited by: rsar001 on Feb 11, 2011 9:10 PM

    Thank you all for the great ideas, you have been extremely helpful. Here is what we did and were able to resolve the query.
    We started with Etbin's idea to create a table from the ext table so that we can index and reference easier than an external table, so we did the following:
    SQL> create table ext_person as select P_ID,ED_ID,trim(F_NAME) fst_name,trim(L_NAME) lst_name from EXT_NAMES_CSV;
    Table created.
    SQL> desc ext_person
    Name                                                                                Null?    Type
    P_ID                                                                                         NUMBER
    ED_ID                                                                                        NUMBER
    FST_NAME                                                                                     VARCHAR2(300)
    LST_NAME                                                                                     VARCHAR2(300)
    SQL> select count(*) from ext_person;
      COUNT(*)
         93383
    SQL> CREATE INDEX EXT_PERSON_ED ON ext_person (P_ID,ED_ID) TABLESPACE D_INDEX;
    Index created.
    SQL> exec dbms_stats.gather_index_stats(ownname=>'APPD', indname=>'EXT_PERSON_ED',partname=> NULL , estimate_percent=> 30 );
    PL/SQL procedure successfully completed.We had a look at the plan with the original SQL query that we had:
    SQL> explain plan for update temp_person
      2  set first_name = (select fst_name
      3                    from ext_person
      4                               where temp_person.PERSON_ID=ext_person.p_id
      5                               and   temp_person.DISTRICT_ID=ext_person.ed_id);
    Explained.
    SQL> @?/rdbms/admin/utlxpls.sql
    PLAN_TABLE_OUTPUT
    Plan hash value: 1236196514
    | Id  | Operation                    | Name           | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | UPDATE STATEMENT             |                | 93383 |  1550K|   186K (50)| 00:37:24 |
    |   1 |  UPDATE                      | TEMP_PERSON    |       |       |            |          |
    |   2 |   TABLE ACCESS FULL          | TEMP_PERSON    | 93383 |  1550K|   191   (1)| 00:00:03 |
    |   3 |   TABLE ACCESS BY INDEX ROWID| EXTT_PERSON    |     9 |  1602 |     1   (0)| 00:00:01 |
    |*  4 |    INDEX RANGE SCAN          | EXT_PERSON_ED  |     1 |       |     1   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       4 - access("EXT_PERSON"."P_ID"=:B1 AND "RS_PERSON"."ED_ID"=:B2)
    Note
       - dynamic sampling used for this statement (level=2)
    20 rows selected.As you can see the time has dropped to 37min (from 6 hrs). Then we decided to change the SQL query and use donisback's suggestion (using MERGE); we explained the plan for teh new query and here is the results:
    SQL> explain plan for MERGE INTO temp_person t
      2  USING (SELECT fst_name ,p_id,ed_id
      3  FROM  ext_person) ext
      4  ON (ext.p_id=t.person_id AND ext.ed_id=t.district_id)
      5  WHEN MATCHED THEN
      6  UPDATE set t.first_name=ext.fst_name;
    Explained.
    SQL> @?/rdbms/admin/utlxpls.sql
    PLAN_TABLE_OUTPUT
    Plan hash value: 2192307910
    | Id  | Operation            | Name         | Rows  | Bytes |TempSpc| Cost (%CPU)| Time     |
    |   0 | MERGE STATEMENT      |              | 92307 |    14M|       |  1417   (1)| 00:00:17 |
    |   1 |  MERGE               | TEMP_PERSON  |       |       |       |            |          |
    |   2 |   VIEW               |              |       |       |       |            |          |
    |*  3 |    HASH JOIN         |              | 92307 |    20M|  6384K|  1417   (1)| 00:00:17 |
    |   4 |     TABLE ACCESS FULL| TEMP_PERSON  | 93383 |  5289K|       |   192   (2)| 00:00:03 |
    |   5 |     TABLE ACCESS FULL| EXT_PERSON   | 92307 |    15M|       |    85   (2)| 00:00:02 |
    Predicate Information (identified by operation id):
       3 - access("P_ID"="T"."PERSON_ID" AND "ED_ID"="T"."DISTRICT_ID")
    Note
       - dynamic sampling used for this statement (level=2)
    21 rows selected.As you can see, the update now takes 00:00:17 to run (need to say more?) :)
    Thank you all for your ideas that helped us get to the solution.
    Much appreciated.
    Thanks

  • Update statement takes too long to run

    Hello,
    I am running this simple update statement, but it takes too long to run. It was running for 16 hours and then I cancelled it. It was not even finished. The destination table that I am updating has 2.6 million records, but I am only updating 206K records. If add ROWNUM <20 to the update statement works just fine and updates the right column with the right information. Do you have any ideas what could be wrong in my update statement? I am also using a DB link since CAP.ESS_LOOKUP table resides in different db from the destination table. We are running 11g Oracle Db.
    UPDATE DEV_OCS.DOCMETA IPM
    SET IPM.XIPM_APP_2_17 = (SELECT DISTINCT LKP.DOC_STATUS
    FROM [email protected] LKP
    WHERE LKP.DOC_NUM = IPM.XIPM_APP_2_1 AND
    IPM.XIPMSYS_APP_ID = 2
    WHERE
    IPM.XIPMSYS_APP_ID = 2;
    Thanks,
    Ilya

    matthew_morris wrote:
    In the first SQL, the SELECT against the remote table was a correlated subquery. the 'WHERE LKP.DOC_NUM = IPM.XIPM_APP_2_1 AND IPM.XIPMSYS_APP_ID = 2" means that the subquery had to run once for each row of DEV_OCS.DOCMETA being evaluated. This might have meant thousands of iterations, meaning a great deal of network traffic (not to mention each performing a DISTINCT operation). Queries where the data is split between two or more databases are much more expensive than queries using only tables in a single database.Sorry to disappoint you again, but with clause by itself doesn't prevent from "subquery had to run once for each row of DEV_OCS.DOCMETA being evaluated". For example:
    {code}
    SQL> set linesize 132
    SQL> explain plan for
    2 update emp e
    3 set deptno = (select t.deptno from dept@sol10 t where e.deptno = t.deptno)
    4 /
    Explained.
    SQL> @?\rdbms\admin\utlxpls
    PLAN_TABLE_OUTPUT
    Plan hash value: 3247731149
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time | Inst |IN-OUT|
    | 0 | UPDATE STATEMENT | | 14 | 42 | 17 (83)| 00:00:01 | | |
    | 1 | UPDATE | EMP | | | | | | |
    | 2 | TABLE ACCESS FULL| EMP | 14 | 42 | 3 (0)| 00:00:01 | | |
    | 3 | REMOTE | DEPT | 1 | 13 | 0 (0)| 00:00:01 | SOL10 | R->S |
    PLAN_TABLE_OUTPUT
    Remote SQL Information (identified by operation id):
    3 - SELECT "DEPTNO" FROM "DEPT" "T" WHERE "DEPTNO"=:1 (accessing 'SOL10' )
    16 rows selected.
    SQL> explain plan for
    2 update emp e
    3 set deptno = (with t as (select * from dept@sol10) select t.deptno from t where e.deptno = t.deptno)
    4 /
    Explained.
    SQL> @?\rdbms\admin\utlxpls
    PLAN_TABLE_OUTPUT
    Plan hash value: 3247731149
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time | Inst |IN-OUT|
    | 0 | UPDATE STATEMENT | | 14 | 42 | 17 (83)| 00:00:01 | | |
    | 1 | UPDATE | EMP | | | | | | |
    | 2 | TABLE ACCESS FULL| EMP | 14 | 42 | 3 (0)| 00:00:01 | | |
    | 3 | REMOTE | DEPT | 1 | 13 | 0 (0)| 00:00:01 | SOL10 | R->S |
    PLAN_TABLE_OUTPUT
    Remote SQL Information (identified by operation id):
    3 - SELECT "DEPTNO" FROM "DEPT" "DEPT" WHERE "DEPTNO"=:1 (accessing 'SOL10' )
    16 rows selected.
    SQL>
    {code}
    As you can see, WITH clause by itself guaranties nothing. We must force optimizer to materialize it:
    {code}
    SQL> explain plan for
    2 update emp e
    3 set deptno = (with t as (select /*+ materialize */ * from dept@sol10) select t.deptno from t where e.deptno = t.deptno
    4 /
    Explained.
    SQL> @?\rdbms\admin\utlxpls
    PLAN_TABLE_OUTPUT
    Plan hash value: 3568118945
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time | Inst |IN-OUT|
    | 0 | UPDATE STATEMENT | | 14 | 42 | 87 (17)| 00:00:02 | | |
    | 1 | UPDATE | EMP | | | | | | |
    | 2 | TABLE ACCESS FULL | EMP | 14 | 42 | 3 (0)| 00:00:01 | | |
    | 3 | TEMP TABLE TRANSFORMATION | | | | | | | |
    | 4 | LOAD AS SELECT | SYS_TEMP_0FD9D6603_1CEEEBC | | | | | | |
    | 5 | REMOTE | DEPT | 4 | 80 | 3 (0)| 00:00:01 | SOL10 | R->S |
    PLAN_TABLE_OUTPUT
    |* 6 | VIEW | | 4 | 52 | 2 (0)| 00:00:01 | | |
    | 7 | TABLE ACCESS FULL | SYS_TEMP_0FD9D6603_1CEEEBC | 4 | 80 | 2 (0)| 00:00:01 | | |
    Predicate Information (identified by operation id):
    6 - filter("T"."DEPTNO"=:B1)
    Remote SQL Information (identified by operation id):
    PLAN_TABLE_OUTPUT
    5 - SELECT "DEPTNO","DNAME","LOC" FROM "DEPT" "DEPT" (accessing 'SOL10' )
    25 rows selected.
    SQL>
    {code}
    I do know hint materialize is not documented, but I don't know any other way besides splitting statement in two to materialize it.
    SY.

  • Update statement taking too long to execute

    Hi All,
    I'm trying to run this update statement. But its taking too long to execute.
        UPDATE ops_forecast_extract b SET position_id = (SELECT a.row_id
            FROM s_postn a
            WHERE UPPER(a.desc_text) = UPPER(TRIM(B.POSITION_NAME)))
            WHERE position_level = 7
            AND b.am_id IS NULL;
            SELECT COUNT(*) FROM S_POSTN;
            214665
            SELECT COUNT(*) FROM ops_forecast_extract;
            49366
    SELECT count(*)
            FROM s_postn a, ops_forecast_extract b
            WHERE UPPER(a.desc_text) = UPPER(TRIM(B.POSITION_NAME));
    575What could be the reason for update statement to execute so long?
    Thanks

    polasa wrote:
    Hi All,
    I'm trying to run this update statement. But its taking too long to execute.
    What could be the reason for update statement to execute so long?You haven't said what "too long" means, but a simple reason could be that the scalar subquery on "s_postn" is using a full table scan for each execution. Potentially this subquery gets executed for each row of the "ops_forecast_extract" table that satisfies your filter predicates. "Potentially" because of the cunning "filter/subquery optimization" of the Oracle runtime engine that attempts to cache the results of already executed instances of the subquery. Since the in-memory hash table that holds these cached results is of limited size, the optimization algorithm depends on the sort order of the data and could suffer from hash collisions it's unpredictable how well this optimization works in your particular case.
    You might want to check the execution plan, it should tell you at least how Oracle is going to execute the scalar subquery (it doesn't tell you anything about this "filter/subquery optimization" feature).
    Generic instructions how to generate a useful explain plan output and how to post it here follow:
    Could you please post an properly formatted explain plan output using DBMS_XPLAN.DISPLAY including the "Predicate Information" section below the plan to provide more details regarding your statement. Please use the {noformat}[{noformat}code{noformat}]{noformat} tag before and {noformat}[{noformat}/code{noformat}]{noformat} tag after or the {noformat}{{noformat}code{noformat}}{noformat} tag before and after to enhance readability of the output provided:
    In SQL*Plus:
    SET LINESIZE 130
    EXPLAIN PLAN FOR <your statement>;
    SELECT * FROM TABLE(DBMS_XPLAN.DISPLAY);Note that the package DBMS_XPLAN.DISPLAY is only available from 9i on.
    In 9i and above, if the "Predicate Information" section is missing from the DBMS_XPLAN.DISPLAY output but you get instead the message "Plan table is old version" then you need to re-create your plan table using the server side script "$ORACLE_HOME/rdbms/admin/utlxplan.sql".
    In previous versions you could run the following in SQL*Plus (on the server) instead:
    @?/rdbms/admin/utlxplsA different approach in SQL*Plus:
    SET AUTOTRACE ON EXPLAIN
    <run your statement>;will also show the execution plan.
    In order to get a better understanding where your statement spends the time you might want to turn on SQL trace as described here:
    When your query takes too long ...
    and post the "tkprof" output here, too.
    Regards,
    Randolf
    Oracle related stuff blog:
    http://oracle-randolf.blogspot.com/
    SQLTools++ for Oracle (Open source Oracle GUI for Windows):
    http://www.sqltools-plusplus.org:7676/
    http://sourceforge.net/projects/sqlt-pp/

  • Average processing time too long in workload monitoring (ST03N)

    Hi expert,
    I wander why average processing time too long in workload monitoring (ST03N)
    for example
    avg.response time : 4700 (ms)
    avg.processing time : 4200 (ms)
    avg.CPU time : 300 (ms)
    avg.DB time : 200(ms)
    OS (CPU and memory ) status is nomal.
    please tell me why processing time too long and what do I check more.
    thanks

    Hi
    Processing Time is the total time taken i.e. Time when user clicks on submit button to send data to SAP System till the time SAP System processes everything and shows the updated screen back to the user. It includes Network Time also.
    Average Processing Time is the Processing Time per Dialog Step
    In ST03N workload Monitor, there are different Task Type like Dialog, Background, RFC, Update,..........
    Could you be more specific where do you find the Average Processing Time high ? I mean for which Task Type.
    If you have users complaining for Performance problem check their Network Time also.
    Let us know, if this information was helpful to you.
    with regards,
    Parin Hariyani

  • SQL Statement taking too long to get the data

    Hi,
    There are over 2500 records in a table and when retrieve all using ' SELECT * From Table' it is taking too long to get the data. ie .. 4.3 secs.
    Is there any possible way to shorten the process time.
    Thanks

    Hi Patrick,
    Here is the sql statement and table desc.
    ID     Number
    SN     Varchar2(12)
    FN     Varchar2(30)
    LN     Varchar2(30)
    By     Varchar(255)
    Dt     Date(7)
    Add     Varchar2(50)
    Add1     Varchar2(30)
    Cty     Varchar2(30)
    Stt     Varchar2(2)
    Zip     Varchar2(12)
    Ph     Varchar2(15)
    Email     Varchar2(30)
    ORgId     Number
    Act     Varchar2(3)     
    select A."FN" || '' '' || A."LN" || '' ('' || A."SN" || '')'' "Name",
    A."By", A."Dt",
    A."Add" || ''
    '' || A."Cty" || '', '' || A."Stt" || '' '' || A."Zip" "Location",
    A."Ph", A."Email", A."ORgId", A."ID",
    A."SN" "OSN", A."Act"
    from "TBL_OPTRS" A where A."ID" <> 0 ';
    I'm displaying all rows in a report.
    if I use 'select * from TBL_OPTRS' , this also takes 4.3 to 4.6 secs.
    Thanks.

  • Elapsed Timer Too Long?

    Is there a limit to the duration I can input into an elapsed timer VI? In my program the user can specify a duration of up to 24 hrs. My program is supposed to hold current conditions for the specified durations. When the user sets it 1, 2, 3 or four hours it holds the current conditions fine. When they set it to 24 hours the current conditions are not held for 24 hours. Is this too long of a duration for this VI (86,400 seconds)?
    thanks.

    rchaoua wrote:
    The question regarding maximum duration a elapsed time VI has still not been answered.
    You can convert the "elapsed time" express VI to a plain VI by "right-click...open front panel" and inspect the code yourself. There are no secrets! You can see that it converts the timestamp to DBL and operates on the lower precision DBL values. This should be sufficient for much more than 24 hours, so something else is wrong in your case.
    Can you attach an example that shows the incorrect behavior (=premature elapsation?).
    rchaoua wrote:
    But I have another question; Can you have multiple elapsed time VI in a single piece of code and not have it affect the other elapsed VIs?  I have three elapsed timers in my code (in different panels) and get the sneaking supsicion that they are not behaving nicely with each other. But I am not sure. I was hoping someone has expereince with this scenario and could shed some light.
    The crucial subVIs are reentrant and thus there should be no interaction between multiple instances of the "elapsed time" express VI. Each will keep it's own state. Why do you suspect that they interact?
    LabVIEW Champion . Do more with less code and in less time .

  • Profile WS run time too long

    My first test run (in dev) of the custom Java PWS against a SQL server database seems to be taking too long. Its seems to be taking appx. 1 second or more per user. This will be a problem when processing around 20,000 users. We can't have a PWS running for 4-6 hours everynight.
    I'd really appreciate any tips to optimize this for faster performance. I am opening DB Connection and running 2 queries for each users and clsoing the statement and connection. Can I do this another way so I open the DB connection(and maybe statement) for the Profile Source only once and release the connection etc. after the whole job is done or in case of errors.
    Please help as this could be a huge bottleneck for us!
    Thanks.
    Vanita
    Staples

    Hi Akash,
    Thanks for your quick reply. Unfortunately I don't have any signature field to limit the profile sync by. I am currently running the PWs in ToMCAT 4.1.30. We plan to deploy on WAS in dev. I am not sure if there is connection pooling being done by Tomcat in 4.1.30 version. Do you have any information on that? I am doing db connect and 2 property queries in the getuserproperties method.
    1. Would the connection pooling (if done in WAS) make any difference to the run time?
    2. Would opening the connection in initialize() method versus GetUserProperties() make any difference?
    As always, thanks for your help.
    Vanita
    ------- Akash Jain wrote on 3/1/05 1:25 PM -------Hi Vanita,On recent hardware, you should be able to perform an initial profile sync at at a rate of ~10/second. This means you should be able to perform your 20k users profile sync in under an hour. Resyncs should be much faster if you use a signature attribute.
    I'm going to assume you're hitting some database backend with a table structure like the following:Users Table String UserGUID Date LastModified
    Properties Table int PropID String UserGUID String PropValue
    You have your users keyed off a unique name - a GUID in this example - and properties in a seperate table keyed off PropID and GUID.
    Lets review the protocol and each step:a) initialize() - sends the parameters of the profile sync to the PWS, this is a good place to do a single query to your database in order to cache all user unique names and signatures (LastModified dates) in a HashTable. This will make re-syncs much faster since subsequent AttachToUser() and GetUserSignature() calls will be derived from this HashTable.b) attachToUser() - in this call you can simply lookup your user record against the HashTable created in Initialize(). If no entry is returned, then throw a NoSuchUserException. If a user does exist continue.c) getUserSignature() - again use the HashTable created in Initialize() to lookup the signature for this user. return it as a String.d) getUserProperties() - if called, this means the signature you sent back in step (c) has changed since the last profile sync. you now want to make a call to your properties table (a single DB call) to load all the property values for the user. return these as a UserPropertyInfo object.
    During an initital sync, you will always get to step (d) above. During re-syncs, assuming there is low churn, I'd say a max of 1% of your calls will get to step (d) and thus the re-sync should be an order of magnitude faster in most cases.
    With respect to database connections - if you are opening and closing a connection for each user, this is pretty poor with respect to performance. Your best bet in Java is to use a single connection (this is a single threaded process) which is setup in initialize(). In the shutdown() method, close this connection.
    I hope this helps, the combination of using a single connection, using the signature attribute and caching all the users unique names and signatures in one call at the start of the profile sync should drastically increase performance.
    Thanks,Akash

  • Update is taking too long to install

    I'm in the process of installing an update to iTunes (11.0.4) and what I think is a security update on my MBP. However, the installation process is taking too long. First, it was stuck for about 2 hours on "registering updated components", and now (for the last hour) I'm seeing a blue screen with the spinning gear. Should I just keep waiting? Has anyone had a similar experience? 

    zornie,
    Try going to safari>preferences, general tab, and setting the "save downloaded files to your desktop (if desktop is not in the drop down menu, click other then select desktop from the left side bar). This will download the update to your dektop as a stand alone installer. Also if for some reason the download is interupted you can click the download icon on the desktop and it will resume where it left off. When the download is complete double click the icon/installer package and follow the prompts. Do them one at a time. (Thsi is not using software update, which can some times get corrupted during DL, especially if your are doing it wireless)
    Security update:
    http://support.apple.com/kb/DL1660
    iTunes:
    http://www.apple.com/itunes/download/
    Hope this helps

  • NI7358 Acceleration time too long

    Hi all,
    First, let me introduces briefly our application for your better understanding before touch the problem.
    We're using NI7358 to control 8 servo motors(Mitsubishi HC-KFS13, 100W, 3000RPM, 131072Counts/Rev) for pick and place. The amplifier is also Mitsubishi one MELSERVO-J2M-P8A. From view of NI7358 all 8 axes are open-loop stepper while amplifier will close-loop control the servo motors. One limitation here is this amplifier only can accept 200KPPS pulse rate input, so we set amplifier internal gear ratio to 4096/125 to achieve its maximum speed.
    GearRatio=(3000rpm*131072count/60)/(200*1000counts/sec)
    Then based on this ratio we can know the resolution of "stepper" in view of NI7358 is 4000RPM. (131072*125/4096)
    Now our problem is:
    The system requires these motors move in short distance(1200steps for NI7358) in the shortest time. So we want to shorten the acceleration/deceleration time as possible as we can. As we test the same application before using Keyence PLC as pulse generator, it can meet our target so we expect NI also can.
    During testing using MAX under 1-D interactive, when I set Velocity=3000rpm, Accel/Decel=20000rps/s, target=1200steps, I can get the best result around 1040rpm(peak velocity) with Accel/Decel time around 50ms. The time is too long for us. If I set increase Accel/Deceleration to its maximum* NI7358 will stop pulse output at all(don't know why but suspect it's because of short distance)
    * Maximum Accel/Deceleration(RPS/s) = Amax X (1/Ts) X (1/Ts) X (1/R) (refer to the description of "flex_load_rpsps")
       for our case: Amax=32counts/sample; Ts=0.25ms; R=4000steps/Rev, result is 128000(RPS/s)
    Questions:
    1. Is Amax=32counts/sample hardware fixed?
    2. Any reason can cause NI7358 stop output pulse when I set maximum acceleration/decelration?
    3. Does NI provide time-based profile instead of velocity one so we can set acceleration/deceleration time directly?
    4. Can I use direct pulse output mode just like a pulse generator instead of current profile output?
    5. Under open-loop mode, if I change PID & acceleration feedforward gain, do these will affect the output pulse performance?
    6. If we change the control circuit from open-loop stepper to close-loop servo system, does it can solve our problem?
    7. How does the trajectory generator produces the count/steps output during the acceleration? I create an excel file(attached below) try to draw out the relationship between the sample, velocity and acceleration, anyone can check for me it's correct or not?
    My formula is here:
    Position2 = Position1 + Velocity2*(Sample2-Sample1)
    Velocity2  = Velocity1 + Acceleration*(Sample2-Sample1)
    All comments are appreciated!
    Attachments:
    Profile Study.xls ‏28 KB

    BO has no controls over VPN or network bamdwith, it takes what OS and DB drivers give it.
    You'll need to identify at which point of the workflow to refresh the report the slowness is occuring.
    You'll need to enable webi tracing and see how long the actual refresh take.
    You'll need to enable network traces to see where in the network the slowdown is occuring.

  • Download time too long

    Why is it taking too long to download a song?  Im talking HOURS if it even happens!

    Could you download the graphics dynamically as they are required?
    So, the air app would be the master 'app' and those additional 
    graphics are loaded from your server only when required? I'd bet this 
    would be the way you'd build it if the app were purely based online 
    and not via air?
    Dave Cates
    Managing Director
    Redemption Media Ltd
    www.redemptionmedia.co.uk
    Interactive internet, media & communications solutions.
    Sent from my iPhone

  • Update statement taking too much time

    Hi All,
    I defined a cursor in which there are almost 94 lakh records and i want to update few fields of a table from another table which takes data over a dblink.
    my cursor is
    cursor c_tpo is
          select t.loc_loc_id,
                 t.till_no,
                 t.transaction_date,
                 t.sales_audit_txn_no,
                 t.third_party_order_ref
            from dwt_om_email_addresses t
           where (t.merc_loaded_date_time is null or t.ws_ordered_date_time is null or
                 t.copos_ordered_date_time is null)
    -- And for loop is
    for i in c_tpo loop
        select count(*)
          into ln_count
          from web_sales@nrstp
         where dmw_order = i.third_party_order_ref;
        if ln_count = 1 then
          select date_loaded, order_date, order_date, pkt_ctrl_nbr
            into ld_merc_load_date,
                 ld_ws_order_date,
                 ld_cop_order_date,
                 lc_pkt_ctrl_nbr
            from web_sales@nrstp
           where dmw_order = i.third_party_order_ref;
          update dwt_om_email_addresses t
             set merc_loaded_date_time   = ld_merc_load_date,
                 ws_ordered_date_time    = ld_ws_order_date,
                 copos_ordered_date_time = ld_cop_order_date
           where t.loc_loc_id = i.loc_loc_id
             and t.till_no = i.till_no
             and t.transaction_date = i.transaction_date
             and t.sales_audit_txn_no = i.sales_audit_txn_no;
      end if;
        ln_count          := 0;
        ln_count_pkt      := 0;
        ld_merc_load_date := null;
        ld_ws_order_date  := null;
        ld_cop_order_date := null;
        ld_pkms_date_time := null;
        lc_pkt_ctrl_nbr   := null;
      end loop;
      -----------The ln_count is used to avoid the exception no_data_found and too_many_rows. How do I speed up the process as it takes 1 hr to update 500 records.
    What would be the best approach to achieve this in less time.
    Thanks

    if you check the your condition some where you have metioned
    dmw_order = t.third_party_order_ref
    if both of these are unique keys or primary key then there is no problem
    but if for there are multiple values edit the query by using
    rownum <= 1
    update dwt_om_email_addresses t
    set (merc_loaded_date_time,ws_ordered_date_time,copos_ordered_date_time) =(select date_loaded, order_date, order_date, pkt_ctrl_nbr from( select date_loaded, order_date, order_date, pkt_ctrl_nbr
                                                                                 from web_sales@nrstp
                                                                                  where dmw_order = t.third_party_order_ref)
    where rownum<=1)
    where exists  (select '1'
                  from web_sales@nrstp i
                  where t.loc_loc_id = i.loc_loc_id
                  and t.till_no = i.till_no
                  and t.transaction_date = i.transaction_date
                  and t.sales_audit_txn_no = i.sales_audit_txn_no
                  and (t.merc_loaded_date_time is null or t.ws_ordered_date_time is null or
                 t.copos_ordered_date_time is null));and for exists clause there is no problem
    Edited by: 810345 on Jun 10, 2011 4:07 PM
    Edited by: 810345 on Jun 10, 2011 4:08 PM

  • ABAP select statements takes too long

    Hi,
    I have a select statement as shown below.
    SELECT * FROM BSEG INTO TABLE ITAB_BSEG
                         WHERE  BUKRS = CO_CODE
                         AND    BELNR IN W_DOCNO
                         AND    GJAHR = THISYEAR
                         AND    AUGBL NE SPACE.
    This select statement runs fine in all of R/3 systems except for 1. The problem that is shown with this particular system is that the query takes very long ( up to an hour for if W_DOCNO consists of 5 entries). Sometimes, before the query can complete, an ABAP runtime error is encountered as shown below:
    <b>Database error text........: "ORA-01555: snapshot too old: rollback segment   
    number 7 with name "PRS_5" too small?"                                       
    Internal call code.........: "[RSQL/FTCH/BSEG ]"                              
    Please check the entries in the system log (Transaction SM21).  </b> 
    Please help me on this issue. However, do not give me suggestions about selecting from a smaller table (bsik, bsak) as my situation does not permit it.
    I will reward points.

    dont use select * ....
    instead u declare ur itab with the required fields and then in select refer to the fields in the select .
    data : begin of itab,
             f1
             f2
             f3
             f4
             end of itab.
    select f1 f2 f3 f4 ..
         into table itab
    from bseg where ...
    . this improves the performance .
    select * is not advised .
    regards,
    vijay

  • Windows 8.1 first boot time too long

    After Sysprep, windows 8.1 tooks long time to boot, is it normal?
    N.A.Malik

    Hi,
    Generally, when you sysprep an image and deploy it to other devices, it would prepare the device settings, registry update and other computer configuration. This phase would take some time. We call it as specialize and OOBE pass.
    Here we have a metric to assess the first boot time, you can take a look at this guide:
    First Boot Performance
    http://msdn.microsoft.com/en-us/library/windows/hardware/jj130825.aspx
    Alex Zhao
    TechNet Community Support

  • Macbook boot time too long

    Hello, boot time on my MacBook seems to be very long. Can you please take a look at my bootchart and tell me what do you think?
    And also the gnome login is very slow. More precisely, it takes too much time for the desktop to appear after logging in.
    Here is the bootchart:
    http://i49.tinypic.com/5yrk11.png
    Last edited by exapplegeek (2012-07-08 22:00:21)

    exapplegeek, your thread title is far too general. Please rephrase it such like: "Macbook boot time problem" as soon as possible.
    And leave the "[BOOTCHART]" part out.

Maybe you are looking for

  • Down Payment Order

    When tried to post the downpayment order to FI. Following message is received. You entered a tax code for a down payment or down payment request which is defined as an EC tax code.  This is not permitted, since this posting is not subject to EC tax. 

  • Reading a String with dynamic format.

    Hello, I'm learning Java and I'm relatively new to the scene. I decided to make a small free utility with the hope that it will help somebody. The utility is a Game Server (Call of Duty 4) connector that will help Administrators to connect to the ser

  • Transfer a char type varialble value to NUMC type variable

    Hi all, I have a BSEG-PROJK  type NUMC(8) variable and one xyz type CHAR(50)  variable. I have to pass value from xyz to BSEG-PROJK my e.g is that while I am passing xyz= 'M-000064' BSEG-PROJK  = xyz then the value of BSEG-PROJK is 00000064 but i nee

  • Include jsp versus seperate class/bean versus taglib

    Hi friends, I am working on Servlet/JSP project. There are large no. of jsp pages. A common set of methods get repeated in all the jsp pages. Now my concen is to separate out the methods in a file and then use it in jsp page. Kindly suggest me the ri

  • My wifi don't work

    Dear Mrs          my computer is the least type :macbook air MD761ZP/A  .   Nowing, my wifi don't work :  My wifi performance at work sometimes connected, sometimes disconnected great impact speed.    I require that the least wifi update 1.0 !