Macbook taking too long o start

why is my macbook pro taking too long (interminable) to start, and is still showing the gray screen with the apple logo?

Hi gersedd!
I have an article for you here that seems to describe your issue and can provide you with some troubleshooting steps for this issue:
Mac OS X: Gray screen appears during startup
http://support.apple.com/kb/ts2570
Thanks for being a part of the Apple Support Communities!
Cheers,
Braden

Similar Messages

  • HT4736 Taking too long (2 minutes) to locate and download photos into Photo '11 9.5 (902.7 build running on an older Intel based MacBook Pro with iPhoto libraries on a USB2 External HD).  I checked and repaired HD permissions. Any ideas?

    Regarding iPhoto '11 9.5 (902.7 build running on an older Intel based MacBook Pro with iPhoto libraries on a USB2 External HD).  I am dealing with iPhoto taking too long to download photos.  Specifically, I rechecked and repaired HD permissions. I am running the most current software my five year old Intel MacBook Pro can run.   What happens is that when I connect an external SD card, or my iPhone, the new version of iPhoto takes up to two full minutes to fully acknowledge the device. Then locate new photos and be ready to download them to my external HD.  I am kind of concerned about this.  This has never happened before. 
    I take 20,000 photos a year.  I really don't want to lose any.  Or is there something I am doing wrong?  Or need to be aware of?  Any experienced suggestions would be appreciated.  Thanks.  Have a great day.
    PS.... The cameras I use are Canon SX-30, Nikon D3100, and my iPhone 4S.  Thanks again for your assistance.

    Hello Old Toad.... Those sound like great ideas. 
    I thought I checked and repaired disk permissions on my main boot HD.  That boot disk is Mac OS Extended (Journaled)  Capacity 749.3 GB.  Available 562.53 GB.   BUT.... now that I think of it.... the Seagate external HD with USB2 interface is: Mac OS Extended (Journaled), Capacity 639.79 GB, Available 36.2 GB with my latest iPhoto Library 517.37 GB that was already scanned & updated to be read by the latest iPhoto version. 
    I'll try your suggestions tonight as far as double checking 'permissions' and setting up a tiny test library.
    Or maybe it's time to fill up another External HD?
    I appreciate your and anyone else's suggestions to try.
    Have a great day. ~~ David in Rochester NY

  • My macbook is taking too long to turn on

    my macbook is taking too long to turn on.

    Try doing a NVRAM/PRAM reset - after you have done this - go to system preferences - startup disk and select your hard drive as the startup drive and restart your Macbook again.
    Resetting NVRAM / PRAM
    Shut down your Mac.
    Locate the following keys on the keyboard: Command (⌘), Option, P, and R. You will need to hold these keys down simultaneously in step 4.
    Turn on the computer.
    Press and hold the Command-Option-P-R keys before the gray screen appears.
    Hold the keys down until the computer restarts and you hear the startup sound for the second time.
    Release the keys.
    After resetting NVRAM or PRAM, you may need to reconfigure your settings for speaker volume, screen resolution, startup disk selection, and time zone information.

  • AirPlay Screen Mirroring (Mavericks) disconnects frequently with "Feedback taking too long to send" message

    My company has several TVs with AppleTVs (3rd generation units) connected in our conference rooms so we can "Screen Mirror" our Mac laptops via AirPlay during meetings. Many employees have complained that AirPlay Screen Mirroring drops frequently during meetings for no apparent reason.
    In attempts to determine the cause of the issue, I removed the AppleTV units from Wi-Fi and hard-wired them all to the LAN (100Mbps/Full duplex, no switchport errors seen on the Cisco switch). I upgraded the AppleTVs to the latest firmware. I had our AppleTV users ensure they were running MacOS Mavericks with the latest software updates installed. I had the Mac laptops hard-wired into the LAN during meetings in the conference rooms. None of these changes resolved the AirPlay issue.
    I reviewed the MacOS "/var/log/system.log" file from the laptops of several users that reported issues. I found a pattern that seemed to indicate that the "coreaudiod" process reported "Feedback taking too long to send" several times before the AppleTV connection was terminated. Also, from a network trace (using "tcpdump") taken during an unexpected AirPlay Screen Mirroring disconnection, I could see that the Mac laptop sent a TCP FIN packet to the AppleTV unit (this would indicate that the MacOS laptop initiated the closing of the AirPlay connection).
    I have included the relevant log file entries below. Please note that the LAN internal to our company is "solid" and there have been no connectivity issues detected or reported during the times the AirPlay sessions were disconnected.
    I believe I have found a workaround to this issue. By going into "System Preferences", "Sound" and then changing the "Output" device BACK to the "Internal Speakers" (rather than the AirPlay destination), the AirPlay Screen Monitoring connection seems to remain stable.
    My questions are:
    - is anyone else experiencing this type of problem? any other solutions recommended?
    - is there a way to change the AirPlay defaults so that Screen Mirroring only sends the video (not audio)?
    - does anyone know what the log file entries indicate (like, what does "Feedback taking too long to send...." mean)?
    - any fix planned for this issue?
    From: "/var/log/system.log":
    Jan 16 10:50:16 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:16.454404 AM [AirPlay] ### Feedback taking too long to send (1 seconds, 1 total)
    Jan 16 10:50:18 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:18.524517 AM [AirPlay] ### Feedback taking too long to send (4 seconds, 2 total)
    Jan 16 10:50:20 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:20.533639 AM [AirPlay] ### Feedback taking too long to send (6 seconds, 3 total)
    Jan 16 10:50:22 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:22.548168 AM [AirPlay] ### Feedback taking too long to send (8 seconds, 4 total)
    Jan 16 10:50:24 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:24.554522 AM [AirPlay] ### Feedback taking too long to send (10 seconds, 5 total)
    Jan 16 10:50:24 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:24.554809 AM [AirPlay] ### Report network status (3, en0) failed: 1/0x1 kCFHostErrorHostNotFound / kCFStreamErrorSOCKSSubDomainVersionCode / kCFStreamErrorSOCKS5BadResponseAddr / kCFStreamErrorDomainPOSIX / evtNotEnb / siInitSDTblErr / kUSBPending / dsBusError / kStatusIsError / kOTSerialSwOverRunErr / cdevResErr / EPERM
    Jan 16 10:50:26 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:26.545531 AM [AirPlay] ### Feedback taking too long to send (12 seconds, 6 total)
    Jan 16 10:50:28 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:28.559050 AM [AirPlay] ### Feedback taking too long to send (14 seconds, 7 total)
    Jan 16 10:50:30 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:30.628868 AM [AirPlay] ### Feedback taking too long to send (16 seconds, 8 total)
    Jan 16 10:50:32 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:32.655638 AM [AirPlay] ### Feedback taking too long to send (18 seconds, 9 total)
    Jan 16 10:50:34 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:34.641952 AM [AirPlay] ### Feedback taking too long to send (20 seconds, 10 total)
    Jan 16 10:50:36 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:36.659854 AM [AirPlay] ### Feedback taking too long to send (22 seconds, 11 total)
    Jan 16 10:50:38 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:38.653594 AM [AirPlay] ### Feedback taking too long to send (24 seconds, 12 total)
    Jan 16 10:50:40 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:40.659279 AM [AirPlay] ### Feedback taking too long to send (26 seconds, 13 total)
    Jan 16 10:50:42 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:42.745549 AM [AirPlay] ### Feedback taking too long to send (28 seconds, 14 total)
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.532853 AM [AirPlay] ### Endpoint "AppleTV" feedback error: -6722/0xFFFFE5BE kTimeoutErr
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.533151 AM [AirPlay] ### Feedback failed: -6722/0xFFFFE5BE kTimeoutErr
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.533273 AM [AirPlay] ### Error with endpoint "AppleTV": -6722/0xFFFFE5BE kTimeoutErr
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.533427 AM [BonjourBrowser] Reconfirming PTR for AppleTV._airplay._tcp.local. on en0
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.533588 AM [BonjourBrowser] Reconfirming PTR for 9C207BBD8EA1@AppleTV._raop._tcp.local. on en0
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.533839 AM [AirPlay] ### AirPlay report: Network dead for 10+ seconds after 159 seconds, screen, nm "AppleTV", tp WiFi, md AppleTV3,1, sv 190.9, rt 0, fu 0, rssi -54
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.534104 AM [AirPlay] ### Report network status (5, en0) failed: 1/0x1 kCFHostErrorHostNotFound / kCFStreamErrorSOCKSSubDomainVersionCode / kCFStreamErrorSOCKS5BadResponseAddr / kCFStreamErrorDomainPOSIX / evtNotEnb / siInitSDTblErr / kUSBPending / dsBusError / kStatusIsError / kOTSerialSwOverRunErr / cdevResErr / EPERM
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.534315 AM [AirPlay] Deactivating virtual display stream for quiesce
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.543682 AM [AirPlayScreenClient] Stopping session
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.543815 AM [AirPlay] Quiescing endpoint 'AppleTV'
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.543907 AM [AirPlayScreenClient] Stopping session internal
    Jan 16 10:50:44 My-MacBook-Pro.local coreaudiod[161]: 2014-01-16 10:50:44.544218 AM [AirPlayScreenClient] Stopped session internal
    Jan 16 10:50:44 My-MacBook-Pro.local AirPlayUIAgent[985]: 2014-01-16 10:50:44.544266 AM [AirPlayAVSys] ### Quiesce AirPlay
    Jan 16 10:50:44 My-MacBook-Pro.local SystemUIServer[159]: 2014-01-16 10:50:44.544297 AM [AirPlayAVSys] ### Quiesce AirPlay
    Jan 16 10:50:44 My-MacBook-Pro.local SystemUIServer[159]: 2014-01-16 10:50:44.553084 AM [AirPlayAVSys] ### Quiesce AirPlay
    Jan 16 10:50:44 My-MacBook-Pro.local AirPlayUIAgent[985]: 2014-01-16 10:50:44.554904 AM [AirPlayAVSys] ### Quiesce AirPlay
    Jan 16 10:50:44 My-MacBook-Pro.local SystemUIServer[159]: 2014-01-16 10:50:44.557604 AM [AirPlayAVSys] Ignoring route away when AirPlay not current
    Jan 16 10:50:44 My-MacBook-Pro.local AirPlayUIAgent[985]: 2014-01-16 10:50:44.560307 AM [AirPlayAVSys] Ignoring route away when AirPlay not current
    Jan 16 10:50:44 My-MacBook-Pro.local WindowServer[89]: Display 0x04280880: GL mask 0x21; bounds (0, 0)[1920 x 1080], 62 modes available
    Jan 16 10:50:44 My-MacBook-Pro.local WindowServer[89]: GLCompositor: GL renderer id 0x01022727, GL mask 0x0000001f, accelerator 0x00004ccb, unit 0, caps QEX|MIPMAP, vram 2048 MB
    I am happy to provide more information if needed.
    Thank you.
    -Tim

    I'm currently still experiencing this as well. I've confirmed it occurs on 10.9.1, 10.9.2, and 10.9.3 on MacBook Pro Retina's, Mid-2012 and 2013 MacBooks. It happens on multiple ATV's not just one, all are updated to 6.1.1 and a simple reboot seems to fix it temporarily but it does come back. All the ATV's are connecting to the network via Wireless not Ethernet. These are 3rd Gen ATV's but I checked the serial number and these do not match the bad batch of Apple TV's from 2013 that were offered up for Replacement for Apple due to the bad firmware update. None of the computers have the Firewall turned on. Here's the two logs that we always find after the issue occurs (logs are recent, happened this morning):
    5/30/14 8:57:36.017 AM coreaudiod[183]: 2014-05-30 08:57:36.016946 AM [AirPlay] ### Feedback taking too long to send (30 seconds, 17 total)
    5/30/14 8:57:36.332 AM coreaudiod[183]: 2014-05-30 08:57:36.331492 AM [AirPlay] ### Feedback failed: -6723/0xFFFFE5BD kCanceledErr
    The user will get disconnected from Airplay anywhere between 30 seconds to 3 minutes after logging on and can reconnect but then once again get disconnected after the same time period. One interesting thing to note is that when the Feedback Taking Too Long to send error starts occuring and the countdown to disconnect start ticking down to 30, its solely referring to Audio not being sent over the network and Video is working just fine. If I try to play sound I get another log and the sound doesn't play through the speakers. After a reboot, sound works fine and the Feedback Error's do not show up. I've also tried switching to Internal Speakers (since it defaultly switches to Airplay Speakers) after connecting to Airplay and seeing the Feedback timer start in the Console Logs but even after that the log continues to saw its taking too long to send and disconnects in 30 seconds. 
    This issue has been ongoing for months, I've got a ticket logged as far back as January with this occuring but its infrequent enough that we've just restarted and moved on. I'd say its an issue that occurs to about 5%-10% of meetings but that's an entire meeting that doesn't have the ability to Airplay until someone comes down and reboots it.
    I don't often post in this forum but its still an active issue with no resolution, proof that its occuring on other people's systems, and no firmware updates having been released to correct it. It'd be nice to know of any workarounds other than having to buy some lamp timers for each conference room just to get a functional ATV or putting up a sign that says hey if you get disconnected every 3 minutes, reboot the ATV. The whole reason we're using Apple products is for ease of use otherwise I'd put together a much cheaper solution myself. Any help or recommended troubleshooting steps would be fantastic at this point.

  • It is taking too long to login after i installed kaspersky

    MY macbook air was working fine but after i installed kaspersky internet security and restarted it is it taking too long to login.. What do i do?

    Try restarting in Safe Mode: Starting up in Safe Mode
    Try disabling your Login Items: If you think you have incompatible login items
    Both are temporary measures. Either one may allow you to log in so that you can uninstall Kaspersky. If you are unable to log in write back for additional instructions.
    To reboot normally restart your Mac and log in without holding any keys.

  • Database open (recovery) taking too long

    Hi,
    Ive been using your awesome BerkeleyDB Java Edition for a couple of years, and have been very happy with it.
    I am currently facing an issue with trying to open the database after a disk-full issue (which resulted in the database being unable to write, and hence not closed properly).
    While recovery seems to be operating, it has been taking an inordinate amount of time - 16 hours so far. My database has data of around 200GB, which inflated to over 450GB during deletion of entries, hence gobbling up all free space on disk.
    My questions are:
    * Should i continue to wait for recovery?
    * Is there any chance that recovery is looping?
    * Is there an easier way (DBDump?) to extract data from the database without having to perform recovery?
    Some other information that may help:
    * The recovery has decreased the size of the last significant file, and created 3 new files since it started running.
    * I have been monitoring the open files (using lsof), and they change every now and then to other files, though a good amount of its time is spent near the end of the database.
    Thus, i feel like recovery is running normally, just taking too long. Please let me know your opinion.
    A few other things i should mention regarding my issue:
    * The database was, till yesterday, running on bdb java 3.3.75. After running several hours of recovery, i upgraded to 4.1.10 (since i read about a possible recovery looping bug in one of the versions)
    * Once 4.1.10 started recovery, it spat out errors regarding the last 2 files. Only on deleting those 2 files (the last being 0 bytes, the 2nd-last being about 5k) did the recovery start. Note that the older 3.3.75's recovery never complained about those files. I can post the errors here if relevant.
    * Some of the jdb files (about 500 files out of the 47,000 files that make up the database) are 100 MB files, since i had experimented with larger sized files for a few days, then reverted the setting.
    Would any of these above affect a successful recovery?
    My setup is:
    OS:Linux CentOS 5.2, 64-bit, kernel 2.6.18-92.el5
    JVM: Sun Java 1.6.0_20, 64-bit
    Memory: 16 GB RAM, of which 8 GB is allocated to the java process (-Xmx8000M -Xms8000M)
    BDB cache set to use 6GB RAM (envconfig.setCacheSize(6000000000))
    Only the BDB basic API is being used (Environment, database, cursors). We do not use DPL, or HA features.
    Awaiting your kind response,
    Sushant A

    Hi Sushant,
    * Should i continue to wait for recovery?* Is there any chance that recovery is looping?>
    I'm not aware of a bug that would cause recovery to loop, however, you may want to take thread dumps to see if it is progressing. It isn't easy to tell, however, since each phase of recovery is in fact a loop. What you can tell easily from the thread dumps is whether recovery is blocked (completely stopped) for some reason. I don't know of a bug that would cause this, but it's something I would check for.
    Assuming it is not blocked, I suggest that you leave recovery running, and additionally (in parallel) try to obtain some information about your log. While recovery is running you can run the DbPrintLog utility, which does not itself run recovery. I suggest running the following command, which will tell us in general what your log looks like and in particular how far apart the checkpoints are:
    java -jar je-x.y.z.jar DbPrintLog -h <envHome> -S > <output>Please post the output.
    If checkpoints are not running in your application for some reason, or they are running very infrequently, this can cause VERY long recoveries. Unfortunately, you may have such a problem in your app and not be aware of it, until you crash and have to recover. To guard against this sort of thing in the future, you should keep an eye on the checkpoint frequency. EnvironmentStats.getNCheckpoints and getEndOfLog can together be used to tell how much log is written between checkpoints. We will also be able to see this from the DbPrintLog -S output.
    * Is there an easier way (DBDump?) to extract data from the database without having to perform recovery?DbDump normally runs recovery. DbDump with the -r or -R option does not run recovery, but has other drawbacks. With -r, a large amount of memory may be necessary to dump an accurate representation of your data set. If this fails because you run out of memory, -R can be used, but this will dump multiple versions of each record and it will be up to you to interpret the output.
    If regular recovery does not succeed, then DbDump -r is the next thing to try.
    Would any of these above affect a successful recovery?No, I don't believe so.
    --mark                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • When query is taking too long time

    When query is taking too long time,Where and how to start tuning it?
    Here i've listed few things need to be considered,out of my knowledge and understanding
    1.What the sql is waiting for(wait events)
    2.Parameter modification need to be done at system/session level
    3.The query has to be tuned (using hints )
    4.Gathering/deleting statistics
    List out any other things that need to be taken into account?
    Which approach must be followed and on what basis that approach must be considered?

    When query is taking too long time,Where and how to start tuning it?explain plan will be good start . trace also
    Here i've listed few things need to be considered,out of my knowledge and understanding
    1.What the sql is waiting for(wait events)When Oracle executes an SQL statement, it is not constantly executing. Sometimes it has to wait for a specific event to happen befor it can proceed.
    Read
    http://www.adp-gmbh.ch/ora/tuning/event.html
    2.Parameter modification need to be done at system/session levelDepend on parameter , define parameter , trace done on session level for example
    3.The query has to be tuned (using hints )Could be help you but you must know how to use .
    4.Gathering/deleting statisticsDo it in non working hours , it will impact on database performance , but its good
    List out any other things that need to be taken into account?Which account ?
    Which approach must be followed and on what basis that approach must be considered?you could use lot of tools , Trace , AWR

  • SQL Update statement taking too long..

    Hi All,
    I have a simple update statement that goes through a table of 95000 rows that is taking too long to update; here are the details:
    Oracle Version: 11.2.0.1 64bit
    OS: Windows 2008 64bit
    desc temp_person;
    Name                                                                                Null?    Type
    PERSON_ID                                                                           NOT NULL NUMBER(10)
    DISTRICT_ID                                                                     NOT NULL NUMBER(10)
    FIRST_NAME                                                                                   VARCHAR2(60)
    MIDDLE_NAME                                                                                  VARCHAR2(60)
    LAST_NAME                                                                                    VARCHAR2(60)
    BIRTH_DATE                                                                                   DATE
    SIN                                                                                          VARCHAR2(11)
    PARTY_ID                                                                                     NUMBER(10)
    ACTIVE_STATUS                                                                       NOT NULL VARCHAR2(1)
    TAXABLE_FLAG                                                                                 VARCHAR2(1)
    CPP_EXEMPT                                                                                   VARCHAR2(1)
    EVENT_ID                                                                            NOT NULL NUMBER(10)
    USER_INFO_ID                                                                                 NUMBER(10)
    TIMESTAMP                                                                           NOT NULL DATE
    CREATE INDEX tmp_rs_PERSON_ED ON temp_person (PERSON_ID,DISTRICT_ID) TABLESPACE D_INDEX;
    Index created.
    ANALYZE INDEX tmp_PERSON_ED COMPUTE STATISTICS;
    Index analyzed.
    explain plan for update temp_person
      2  set first_name = (select trim(f_name)
      3                    from ext_names_csv
      4                               where temp_person.PERSON_ID=ext_names_csv.p_id
      5                               and   temp_person.DISTRICT_ID=ext_names_csv.ed_id);
    Explained.
    @?/rdbms/admin/utlxpls.sql
    PLAN_TABLE_OUTPUT
    Plan hash value: 3786226716
    | Id  | Operation                   | Name           | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | UPDATE STATEMENT            |                | 82095 |  4649K|  2052K  (4)| 06:50:31 |
    |   1 |  UPDATE                     | TEMP_PERSON    |       |       |            |          |
    |   2 |   TABLE ACCESS FULL         | TEMP_PERSON    | 82095 |  4649K|   191   (1)| 00:00:03 |
    |*  3 |   EXTERNAL TABLE ACCESS FULL| EXT_NAMES_CSV  |     1 |   178 |    24   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       3 - filter("EXT_NAMES_CSV"."P_ID"=:B1 AND "EXT_NAMES_CSV"."ED_ID"=:B2)
    Note
       - dynamic sampling used for this statement (level=2)
    19 rows selected.By the looks of it the update is going to take 6 hrs!!!
    ext_names_csv is an external table that have the same number of rows as the PERSON table.
    ROHO@rohof> desc ext_names_csv
    Name                                                                                Null?    Type
    P_ID                                                                                         NUMBER
    ED_ID                                                                                        NUMBER
    F_NAME                                                                                       VARCHAR2(300)
    L_NAME                                                                                       VARCHAR2(300)Anyone can help diagnose this please.
    Thanks
    Edited by: rsar001 on Feb 11, 2011 9:10 PM

    Thank you all for the great ideas, you have been extremely helpful. Here is what we did and were able to resolve the query.
    We started with Etbin's idea to create a table from the ext table so that we can index and reference easier than an external table, so we did the following:
    SQL> create table ext_person as select P_ID,ED_ID,trim(F_NAME) fst_name,trim(L_NAME) lst_name from EXT_NAMES_CSV;
    Table created.
    SQL> desc ext_person
    Name                                                                                Null?    Type
    P_ID                                                                                         NUMBER
    ED_ID                                                                                        NUMBER
    FST_NAME                                                                                     VARCHAR2(300)
    LST_NAME                                                                                     VARCHAR2(300)
    SQL> select count(*) from ext_person;
      COUNT(*)
         93383
    SQL> CREATE INDEX EXT_PERSON_ED ON ext_person (P_ID,ED_ID) TABLESPACE D_INDEX;
    Index created.
    SQL> exec dbms_stats.gather_index_stats(ownname=>'APPD', indname=>'EXT_PERSON_ED',partname=> NULL , estimate_percent=> 30 );
    PL/SQL procedure successfully completed.We had a look at the plan with the original SQL query that we had:
    SQL> explain plan for update temp_person
      2  set first_name = (select fst_name
      3                    from ext_person
      4                               where temp_person.PERSON_ID=ext_person.p_id
      5                               and   temp_person.DISTRICT_ID=ext_person.ed_id);
    Explained.
    SQL> @?/rdbms/admin/utlxpls.sql
    PLAN_TABLE_OUTPUT
    Plan hash value: 1236196514
    | Id  | Operation                    | Name           | Rows  | Bytes | Cost (%CPU)| Time     |
    |   0 | UPDATE STATEMENT             |                | 93383 |  1550K|   186K (50)| 00:37:24 |
    |   1 |  UPDATE                      | TEMP_PERSON    |       |       |            |          |
    |   2 |   TABLE ACCESS FULL          | TEMP_PERSON    | 93383 |  1550K|   191   (1)| 00:00:03 |
    |   3 |   TABLE ACCESS BY INDEX ROWID| EXTT_PERSON    |     9 |  1602 |     1   (0)| 00:00:01 |
    |*  4 |    INDEX RANGE SCAN          | EXT_PERSON_ED  |     1 |       |     1   (0)| 00:00:01 |
    Predicate Information (identified by operation id):
       4 - access("EXT_PERSON"."P_ID"=:B1 AND "RS_PERSON"."ED_ID"=:B2)
    Note
       - dynamic sampling used for this statement (level=2)
    20 rows selected.As you can see the time has dropped to 37min (from 6 hrs). Then we decided to change the SQL query and use donisback's suggestion (using MERGE); we explained the plan for teh new query and here is the results:
    SQL> explain plan for MERGE INTO temp_person t
      2  USING (SELECT fst_name ,p_id,ed_id
      3  FROM  ext_person) ext
      4  ON (ext.p_id=t.person_id AND ext.ed_id=t.district_id)
      5  WHEN MATCHED THEN
      6  UPDATE set t.first_name=ext.fst_name;
    Explained.
    SQL> @?/rdbms/admin/utlxpls.sql
    PLAN_TABLE_OUTPUT
    Plan hash value: 2192307910
    | Id  | Operation            | Name         | Rows  | Bytes |TempSpc| Cost (%CPU)| Time     |
    |   0 | MERGE STATEMENT      |              | 92307 |    14M|       |  1417   (1)| 00:00:17 |
    |   1 |  MERGE               | TEMP_PERSON  |       |       |       |            |          |
    |   2 |   VIEW               |              |       |       |       |            |          |
    |*  3 |    HASH JOIN         |              | 92307 |    20M|  6384K|  1417   (1)| 00:00:17 |
    |   4 |     TABLE ACCESS FULL| TEMP_PERSON  | 93383 |  5289K|       |   192   (2)| 00:00:03 |
    |   5 |     TABLE ACCESS FULL| EXT_PERSON   | 92307 |    15M|       |    85   (2)| 00:00:02 |
    Predicate Information (identified by operation id):
       3 - access("P_ID"="T"."PERSON_ID" AND "ED_ID"="T"."DISTRICT_ID")
    Note
       - dynamic sampling used for this statement (level=2)
    21 rows selected.As you can see, the update now takes 00:00:17 to run (need to say more?) :)
    Thank you all for your ideas that helped us get to the solution.
    Much appreciated.
    Thanks

  • Iam using 3015 network printer. while iam printing from word it is taking too long to print

    Respected sir,
    I am using 3015 printer.  when ever i give print it is taking too long time nearly 1 minute . there is no problem with the network

    Let's set a static IP address for the printer:
    - Print a Network Config Page from the front of the printer. Note the printer's IP address.
    - Type that IP address into a browser to reveal the printer's internal settings.
    - Choose the Networking tab, then Wireless along the left side, then the IPv4 tab.
    - On this screen you want to set a Manual IP. You need to set an IP address outside the range that the router automatically sets (called the DHCP range). You can find the DHCP range of the router using its internal settings page or in its manual. Use the CD that came with your router or type the router's IP address (ends in .1) into a browser.
    - Use 255.255.255.0 for the subnet (unless you know it is different, if so, use that)
    - Enter your router's IP (on the Network Config Page) for the gateway and first DNS. Leave the second one blank.
    - Click 'Apply'.
    Now, shut down the router and printer, start the router, wait, then start the printer.
    After this you may need to redo 'Add a Printer' using the new IP address.
    Say thanks by clicking "Kudos" "thumbs up" in the post that helped you.
    I am employed by HP

  • Network event is taking too long (100%)

    Hi everybody. We have a 10g DB on Windows. We're using OEM to manage the DB, and it has started to show an alert about database time spend waiting for "Network" event. It arises when we execute one module that updates several tables, that is taking too long. Before we had this app on 8i, also on Windows, and that operation was much more faster than now. The indexes on tables are valid, and I've gathered statistics for the CBO, so I suppose the problem is, as the OEM says, related with network, but I don't know why, because the connection speed is the same tan before, and the two machines are in the same LAN.
    Any ideas??

    Here is the output requested:
    SQL> select * from v$system_event
    2 where event like 'SQL%';
    EVENT TOTAL_WAITS TOTAL_TIMEOUTS TIME_WAITED
    AVERAGE_WAIT TIME_WAITED_MICRO EVENT_ID
    SQL*Net message to client 1159200 0 252
    0 2516408 2067390145
    SQL*Net message to dblink 2234 0 1
    0 5590 3655533736
    SQL*Net more data to client 5753 0 166
    0 1657387 554161347
    EVENT TOTAL_WAITS TOTAL_TIMEOUTS TIME_WAITED
    AVERAGE_WAIT TIME_WAITED_MICRO EVENT_ID
    SQL*Net more data to dblink 12 0 0
    0 548 1958556342
    SQL*Net message from client 1159181 0 218341084
    188 2,1834E+12 1421975091
    SQL*Net more data from client 23299 0 180602
    8 1806015123 3530226808
    EVENT TOTAL_WAITS TOTAL_TIMEOUTS TIME_WAITED
    AVERAGE_WAIT TIME_WAITED_MICRO EVENT_ID
    SQL*Net message from dblink 2234 0 3693
    2 36934861 4093028837
    SQL*Net more data from dblink 4021 0 39
    0 390002 1136294303
    SQL*Net break/reset to client 182986 0 2740
    0 27397165 1963888671
    9 filas seleccionadas.

  • R/3 Extraction taking too long to load data into BW

    HI There,
    I'm trying to extract SAP Standard extractor 0FI_AP_4 into BW, and its taking endless time.
    Even the Extract checker RSA3  is taking too long to execute the data. Dont know why its taking so long to execute.
    Since there in not much data to take such a long time.
    Enhanced the datasource with three fields from BSEG using user exits.
    Is that the reason why its taking too long? Does User Exit slows down the extraction process?
    What measures i should take to quicken the process?
    Thanks for your time
    Vandana

    Thanks for all you replies.
    Please go through the steps I've gone through :
    - Installed the Business Content and its in version 3.5
    - Changed the update rules, Transfer rules and migrated the datasource to BI 7
    - Enhanced the 0FI_AP_3 to include three fields BSEG table
    - Ran RSA3 and the new fields are showing but the loading is quite slow.
    - Commented the code and ran RSA3 and with little difference data is showing up
    - Removed the comments and ran, its fine, though it takes little more time then previous step...but data is showing up
    - Replicated the datasource into BW
    - Created the info package and started the init process (before this deleted the previous stored init process)
    - Data isn't loading and please see the error message below.
    Diagnosis
    The data request was a full update.  In this case, the corresponding table in the source system does not
    contain any data. System Response Info IDoc received with status 8. Procedure Check the data basis in the source system.
    - Checked the transformation between datasource 0FI_AP_4 and Infosource ZFI_AP_4
       and I DID NOT found the three fields which i enhanced from BSEG table in the 0FI_AP_4 datasource.
    - Replicated the datasource 0FI_AP_4 again, but no change.
    Now...I dont know whats happening here.
    When i check the datasource 0FI_AP_4 in RSA6, i can see the three new fields from BSEG.
    When i check RSA3, i can see the data getting populated with the three new fields from BSEG,
    When i check the fields in the datasource 0FI_AP_4 in BW, I can see the three new fields. It shows
    that the connection between BW and R/3 is fine, isn't it?
    Now...Can anyone please suggest me how to go forward from here?
    Thanks for your time
    Vandana

  • Connect by taking too long

    Hello,
    I have a table "a" that has:
    left_id right_id type
    4 5 1
    4 6 1
    4 7 1
    5 9 2
    5 10 2
    5 11 2
    9 13 3
    13 14 4
    10 15 3
    QUERY:
    select left_id, right_id, type from
    a
    connect by left_id = prior right_id
    start with left_id = 4;
    Execution Plan
    Plan hash value: 2739023583
    | Id | Operation | Name | Rows | Bytes | Cost (%CPU)| Time |
    | 0 | SELECT STATEMENT | | 29 | 1131 | 18 (12)| 00:00:01 |
    |* 1 | CONNECT BY WITH FILTERING| | | | | |
    |* 2 | INDEX RANGE SCAN | a_PK | 5 | 65 | 3 (0)| 00:00:01 |
    | 3 | NESTED LOOPS | | 24 | 624 | 13 (0)| 00:00:01 |
    | 4 | CONNECT BY PUMP | | | | | |
    |* 5 | INDEX RANGE SCAN | a_PK | 5 | 65 | 2 (0)| 00:00:01 |
    Predicate Information (identified by operation id):
    1 - access("LEFT_ID"=PRIOR "RIGHT_ID")
    2 - access("LEFT_ID"=4)
    5 - access("LEFT_ID"="connect$_by$_pump$_002"."prior right_id ")
    Is there a way to optimize the query?
    The query is taking too long.
    -Thanks
    Karthik
    Edited by: user3934098 on Nov 14, 2010 1:50 AM

    Here is the detailed explaination:
    Version: oracle 10g R2
    Create table statement:
    CREATE TABLE A
    "LEFT_ID" NUMBER(9,0) NOT NULL ENABLE,
    "RIGHT_ID" NUMBER(9,0) NOT NULL ENABLE,
    "TYPE" NUMBER(9,0) NOT NULL ENABLE,
    CONSTRAINT "A_PK" PRIMARY KEY ("LEFT_ID", "RIGHT_ID", "TYPE") USING INDEX PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT) TABLESPACE "DATA" ENABLE,
    CONSTRAINT "A_FK1" FOREIGN KEY ("TYPE") REFERENCES "B" ("TYPE") ENABLE
    SEGMENT CREATION IMMEDIATE PCTFREE 10 PCTUSED 40 INITRANS 1 MAXTRANS 255 NOCOMPRESS LOGGING STORAGE
    INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT
    TABLESPACE "DATA" ;
    Insert statements:
    INSERT INTO A VALUES(4, 5, 1);
    INSERT INTO A VALUES(4, 6, 1);
    INSERT INTO A VALUES(4, 7, 1);
    INSERT INTO A VALUES(5, 9, 2);
    INSERT INTO A VALUES(5, 10, 2);
    INSERT INTO A VALUES(5, 11, 2);
    INSERT INTO A VALUES(9, 13, 3);
    INSERT INTO A VALUES(13, 14, 4);
    INSERT INTO A VALUES(10, 15, 3);
    INDEXES:
    CREATE UNIQUE INDEX "A_PK" ON "A" ("LEFT_ID", "RIGHT_ID", "TYPE") PCTFREE 10 INITRANS 2 MAXTRANS 255 COMPUTE STATISTICS STORAGE(INITIAL 65536 NEXT 1048576 MINEXTENTS 1 MAXEXTENTS 2147483645 PCTINCREASE 0 FREELISTS 1 FREELIST GROUPS 1 BUFFER_POOL DEFAULT FLASH_CACHE DEFAULT CELL_FLASH_CACHE DEFAULT) TABLESPACE "DATA" ;
    QUERY:
    select left_id, right_id, type from
    a
    connect by left_id = prior right_id
    start with left_id = 4;
    THE table has 951053 rows
    The explain plan is:"
    Execution Plan
    Plan hash value: 2739023583
    Id Operation Name Rows Bytes Cost (%CPU) Time
    0 SELECT STATEMENT 29 1131 18 (12) 00:00:01
    * 1 CONNECT BY WITH FILTERING
    * 2 INDEX RANGE SCAN a_PK 5 65 3 (0) 00:00:01
    3 NESTED LOOPS 24 624 13 (0) 00:00:01
    4 CONNECT BY PUMP
    * 5 INDEX RANGE SCAN a_PK 5 65 2 (0) 00:00:01
    Predicate Information (identified by operation id):
    1 - access("LEFT_ID"=PRIOR "RIGHT_ID")
    2 - access("LEFT_ID"=4)
    5 - access("LEFT_ID"="connect$_by$_pump$_002"."prior right_id ")
    Now is there a way to optimize the query? The query takes about a min to excute. This will be my inner query.
    " Is there any other information that you may need? Am I missing something here? "
    -Thanks
    Karthik
    Edited by: user3934098 on Nov 14, 2010 2:22 AM

  • HT5191 (Windows 7/64bit) iTunes Taking too long to Backup.

    Hi guys/gurus/"i" masters,
    Got a MAJOR issue with itunes. I don't know why this happens on windows but this is one issue i can't solve.
    I'm on an iPhone 4s running ios 5.0.1. I want to backup all my stuff before I bring to the latest firmware (5.1.1)
    I've had to force close the program many times as the backup was taking too long, and it didn't even move a bit (backup bar).
    Can anyone please help me to resolve this issue.
    Appreciate the help very much.
    Regards,
    TJB!

    Start by making sure you have the latest version of iTunes.
    If that doesn't do it, you may be seeing something like what is described in this Apple article -> Apple software on Windows: May see performance issues and blank iTunes Store
    If it's what you're seeing, you can try resetting Winsock as a temporary fix or folllow along with the article.
    1. To do so, please do the following in an administrator user:
    Windows Vista and Windows 7: Right click All Programs and choose Accessories > Command Prompt. Choose to open as an administrator.
    Windows XP: Click Start > Run and type in cmd.
    2. Enter the following command: 
    netsh winsock reset
    3. Restart the computer and test the issue.
    Note:  If the issue returns, third-party software may be reinserting the LSP.

  • Data manager jobs taking too long or hanging

    Hoping someone here can provide some assistance with regard to the 4.2 version. We are specifically using BPC/OutlookSoft 4.2SP4 (and in process of upgrading to BPC7.5). Three server environment - SQL, OLAP and Web.
    Problem: Data manager jobs in each application of production appset with five applications are either taking too long to complete for very small jobs (single entity/single period data copy/clear, under 1000 records) or completely hanging for larger jobs. This has been an issue for the last 7 days. During normal operation, small DM jobs ran in under a minute and large ones taking only a few minutes.
    Failed attempts at resolution thus far:
    1. Processed all applications from the OLAP server
    2. Confirmed issue is specific to our appset and is not present in ApShell
    3. Copied packages from ApShell to application to eliminate package corruption
    4. Windows security updates were applied to all three servers but I assume this would also impact ApShell.
    5. Cleared tblDTSLog history
    6. Rebooted all three servers
    7. Suspected antivirus however, problem persists with antivirus disabled on all three servers.
    Other Observations
    There are several tables in the SQL database named k2import# and several stored procedures named DMU_k2import#. My guess is these did not get removed because I killed the hung up jobs. I'm not sure if their existence is causing any issues.
    To make the long story short, how can I narrow down at which point the jobs are hanging up or what is taking the longest time? I have turned on Debug Script but I don't' have documentation to make sense of all this info. What exactly is happening when I run a Clear package?  At this point, my next step is to run SQL Profiler to get a look into what is going on behind the scenes on the sql server. I also want to rule out the COM+ objects on the web server but not sure where to start.
    Any help is greatly appreciated!!
    Thank you,
    Hitesh

    Hi ,
    The problem seems to be related to database. Do you have any maintenance plan for database?
    It is specific for your appset because each appset has own database.
    I suspect you have to run an sp_updatestats (Update Statistics) for your database and I think the issue with your jobs hang will be solved.
    DMU_K2_importXXX  table are coming from hang imports ...you can delete these tables because it is just growing the size of database and for sure are not used anymore.
    Regards
    Sorin Radulescu

  • Why mac taking too long to wake-up??

    I dont know ahy but my mac is taking too long to wake up after sleep mode. As I open up the lid, the password screen pops up quickly but after I type in the password, it takes too long to open up. Does anyone have a solution for this problem?

    It might have something to do with the SSD, what is the size of your Ram ?? do an extensive test on your RAM which will take a long time. Bad RAM can do wacky things and work fine until you do something specific or push the system hard. Have a look at this articles here, might be helpful: https://www.ewal.net/2012/09/09/slow-wake-for-macbook-pro-retina/

Maybe you are looking for

  • HDMI Video corruption under Mountain Lion

    I have a mid 2011 15 inch Macbook pro.  Upgraded to Mountain Lion.  I use an HDMI TV connection as an external monitor with the laptop closed.  I switch to another work PC on the TV using the source selection.  This causes the Mac to Sleep because no

  • IPhoto can't import rotated movies?

    Folks, I have some movies taken with my digital camera in portrait mode. I found a program to rotate the movies (TransformMovie), but iPhoto doesn't let me import them. It gives an error message. The movies play ok just in Quicktime (or frontrow or w

  • BW testing strategy

    Hi All, I have requirement to create a BW testing strategy for any CRs received. It has to cover general step by step procedure to be followed to test and document all the tests conducted on changes made in BW. As a QA team lead, I have to prepare th

  • Sat Nav - Navigate To Coordinates ?

    Hi, This is my first post on the forum So be nice ! I work for Coca-Cola Enterprises and im trying to find a sat nav app for the BB 8820 that allows me to enter coordinates that i can then navigate too. This is because our vehicle tracking service tr

  • Email password not working

    When I had my broken phone replaced the email password got erased, so I entered it again and it keeps telling me it's the wrong password. I know it's correct because it's the same one I use in Thunderbird. Any thoughts?