Processing File takes too long

I have downloaded a number of albums on iTunes before and in recent times, the processing file is extremely slow. It takes about 95 seconds to process a 14MB file and considering that most of the music I download is around this size it is very frustrating.
My internet speed is normally 54Mbs with a good connection and I have checked that I don't have the V CAST from Verizon so there is no conflict there. I currently have 7GB of music in my library.
Does anyone have any suggestions on what I can do?

I think when it sees that declaration, it tries to actually download the DTD so it can validate the document against it.
So the time delay is the network connection.
There's a simple way to turn this off but I forget what it is.
There are also klugy ways to turn it off but I won't say how so as to not poison the well.

Similar Messages

  • OPM process execution process parameters takes too long time to complete

    PROCESS_PARAMETERS are inserted every 15 min. using gme_api_pub packages. some times it takes too long time to complete the batch ,ie completion of request. it takes about 5-6 hrs long time ,in other time s it takes only 15-20 mins.This happens at regular interval...if anybody can guide me I will be thankful to him/her..
    thanks in advance.
    regds,
    Shailesh

    Generally the slowest part of the process is in the extraction itself...
    Check in your source system and see how long the processes are taking, if there are delays, locks or dumps in the database... If your source is R/3 or ECC transactions like SM37, SM21, ST22 can help monitor this activity...
    Consider running less processes in parallel if you have too many and see some delays in jobs... Also indexing some of the tables in the source system to expedite the extraction, make sure there are no heavy processes or interfaces running in the source system at the same time you're trying to load... Check with your Basis guys for activity peaks and plan accordingly...
    In BW also check in your SM21 for database errors or delays...
    Just some ideas...

  • Copy video files then processing files takes to long

    first i have to copy the video files which takes about 6 hours sometimes more sometimes less then processing the files takes another 6 hours more sometimes less... am i doing something wrong ro how can i speed this up?
    i am copying the files from my mac hd to an external hd... is that the problem?

    i have copies on one external drive and i am copying them to another drive for for permanent storage and when i copy them over it takes long about 6 hrs... after its finished copying it process the files so that takes about the same time as copying them. I even tried doing it straight from my video camera to my mac and after that it has to process the files...i dont understand why it even does this. this never happened with imovie 6 hd or even final cut express it only does this with imovie 8... to me it does not make any sense. i thought things would be faster in imovie 8 but it doesnt seem to be...

  • Swf files take too long to load on server:

    Good Morning Everyone,
      I have a problem. I just uploaded a mostly all flash web site (It has one html page, the main one) and the rest of the buttons all load .swf files.  The problem I am having is that when I click on buttons on most of the pages, the screen will go black for about 15-20 seconds before loading it. The visitor would think it was a broken link and leave the site if that happened!!!
      Here is the link to the site:
      http://www.webwizkid.com/breaker boys flash homepage 2.html
      Try clicking on any of the links from that page, especially the History, Stories and Unit Plan pages.
    PS: This site was designed with  Flash MX and ActionScript II
    My web hosting service is IPOWER.com
    I am wondering if this has something to do with Bandwidth?
    Can anyone tell me what the problem could be?
    Hope to hear from someone soon.
    Sincerely,
    Tom

    How large (file size) are the .swfs that you are loading when a button gets clicked?
    During creation of the .swfs, did you test the .swf with the Bandwidth Profiler? it will tell you approx how long it will take to download at various bandwidth (Internet connection speed).
    http://www.communitymx.com/content/article.cfm?cid=C2A44
    to reduce file size, are you optimizig all the content?
    http://help.adobe.com/en_US/flash/cs/using/WSd60f23110762d6b883b18f10cb1fe1af6-7b23a.html# WSd60f23110762d6b883b18f10cb1fe1af6-7b21a
    Best wishes,
    Adninjastrator

  • I don't know why it takes too long time to sample flat file.

    I don't know why it takes too long time to sample flat file.
    OWB Client 10.1
    While importing a flat file of fixed width ,
    in the screen "Flat File Sample Wizard" shows the text box number of rows with default value 200.
    I want to extend this value to 700,000.
    But, it takes too long time (over 5 hours) to sample it.
    Do you know why it is happend? or How can i fix this problem?
    Thanks in Advance.
    Regards,
    JWS.

    Hello,
    Actually flat file sampling process’ goal is to capture the structure of the file. That’s why initially the sample size is set to 200 lines.
    The question is why you are trying to perform sampling by 700000 rows? Are you expecting some change in structure beyond this mark?
    If so, and you want to capture the fact that your source file is multi – typed, your better prepare small file for sampling outside the OWB.
    Sergey

  • HT1386 copying files in itunes takes too long

    When I drag music files into the iPhone's playlist, the copying takes too long.  I am running 11 and have a 4

    Wait on the update to finish downloading and installing.
    If the device is in recovery mode there is really nothing that can be done besides restoring it.

  • Why does it take too long to open attachments from email account?

    It takes forever to download attachments from my email account. I have tried to do the following tasks explained on the vista forum :http://windowshelp.microsoft.com/Windows/en-US/help/6b046ae9-1434-4423-9303-400ff6fe686b1033.mspx#ESD but none of the possible fixes work.
    After clicking open on the pop up box asking whether i want to open or save the attachement It takes too long to download it. The transfer window stays open showing that it is ready to download but stays at that bhox window. I press cancel and try to open again, if lucky it opens that file, otherwise it takes forever forceing me to cancel. The files very small files most of the time, usually between 50kb so should take seconds.
    I have even tried to save the files but again same process. The transfer box stays open but does not download.
    Any one any ideas.
    Thanx in advance.

    Hello
    It is not easy to say what happen exactly but it must be something with email account provider and their page. For me this case is not some typical Vista problem but you can try to find solution on Microsoft Vista IT Pro forum.
    By the way: do you have alternative mail address by some other provider? Is there the same situation?

  • Why finding replication stream matchpoint takes too long

    hi,
    I am using bdb je 5.0.58 HA(two nodes group,JVM 6G for each node).
    Sometimes, I found bdb node takes too long to restart(about 2 hours).
    When this occurs, I catch the process stack of bdb(jvm process) by jstack.
    After analyzing stack,I found "ReplicaFeederSyncup.findMatchpoint()" taking all the time.
    I want to know why this method takes so much time,and how can I avoid this bad case.
    Thanks.
    帖子经 liang_mic编辑过

    Liang,
    2 hours is indeed a huge amount of time for a node restart. It's hard be sure without doing more detailed analysis of your log as to what may be going wrong, but I do wonder if it is related to the problem you reported in outOfMemory error presents when cleaner occurs [#21786]. Perhaps the best approach is for me to describe in more detail what happens when a replicated node is connecting with a new master, which might give you more insight into what is happening in your case.
    The members of a BDB JE HA replication group share the same logical stream of replicated records, where each record is identified with a virtual log sequence number, or VLSN. In other words, the log record described by VLSN x on any node is the same data record, although it may be stored in a physically different place in the log of each node.
    When a replica in a group connects with a master, it must find a common point, the matchpoint, in that replication stream. There are different situations in which a replica may connect with a master. For example, it may have come up and just joined the group. Another case is when the replica is up already but a new master has been elected for the group. One way or another, the replica wants to find the most recent point in its log, which it has in common with the log of the master. Only certain kinds of log entries, tagged with timestamps, are eligible to be used for such a match, and usually, these are transaction commits and aborts.
    Now, in your previous forum posting, you reported an OOME because of a very large transaction, so this syncup issue at first seems like it might be related. Perhaps your replication nodes need to traverse a great many records, in an incomplete transaction, to find the match point. But the syncup code does not blindly traverse all records, it uses the vlsn index metadata to skip to the optimal locations. In this case, even if the last transaction was very long, and incomplete, it should know where the previous transaction end was, and find that location directly, without having to do a scan.
    As a possible related note, I did wonder if something was unusual about your vlsn index metadata. I did not explain this in outOfMemory error presents when cleaner occurs but I later calculated that the transaction which caused the OOME should only have contained 1500 records. I think that you said that you figured out that you were deleting about 15 million records, and you figured out that it was the vlsn index update transaction which was holding many locks. But because the vlsn index does not record every single record, it should only take about 1,500 metadata records in the vlsn index to cover 15 million application data records. It is still a bug in our code to update that many records in a single transaction, but the OOME was surprising, because 1,500 locks shouldn't be catastrophic.
    There are a number of ways to investigate this further.
    - You may want to try using a SyncupProgress listener described at http://docs.oracle.com/cd/E17277_02/html/java/com/sleepycat/je/rep/SyncupProgress.html to get more information on which part of the syncup process is taking a long time.
    - If that confirms that finding the matchpoint is the problem, we have an unadvertised utility, meant for debugging, to examine the vlsn index. The usage is as follows, and you would use the -dumpVLSN option, and run thsi on the replica node. But this would require our assistance to interpret the results. We would be looking for the records that mention where "sync" points are, and would correlate that to the replica's log, and that might give more information if this is indeed the problem, and why the vlsn index was not acting to optimize the search.
    $ java -jar build/lib/je.jar DbStreamVerify
    usage: java { com.sleepycat.je.rep.utilint.DbStreamVerify | -jar je-<version>.jar DbStreamVerify }
    -h <dir> # environment home directory
    -s <hex> # start file
    -e <hex> # end file
    -verifyStream # check that replication stream is ascending
    -dumpVLSN # scan log file for log entries that make up the VLSN index, don't run verify.
    -dumpRepGroup # scan log file for log entries that make up the rep group db, don't run verify.
    -i # show invisible. If true, print invisible entries when running verify mode.
    -v # verbose

  • Web application deployment takes too long?

    Hi All,
    We have a wls 10.3.5 clustering environment with one admin server and two managered servers separately. When we try to deploy a sizable web application, it takes about 1 hour to finish. It seems that it takes too long to finish the deployment. Here is the output from one of two managerd server system log. Could anyone tell me it is normal or not? If not, how can I improve this?
    Thanks in advance,
    John
    +####<Feb 29, 2012 12:11:03 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535463373> <BEA-149059> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] is transitioning from STATE_NEW to STATE_PREPARED on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:05 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <9baa7a67b5727417:26f76f6c:135ca05cff2:-8000-00000000000000b0> <1330535465664> <BEA-149060> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] successfully transitioned from STATE_NEW to STATE_PREPARED on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466493> <BEA-149059> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] is transitioning from STATE_PREPARED to STATE_ADMIN on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466493> <BEA-149060> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] successfully transitioned from STATE_PREPARED to STATE_ADMIN on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466809> <BEA-149059> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] is transitioning from STATE_ADMIN to STATE_ACTIVE on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466809> <BEA-149060> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] successfully transitioned from STATE_ADMIN to STATE_ACTIVE on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442300> <BEA-320143> <Scheduled 1 data retirement tasks as per configuration.>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320144> <Size based data retirement operation started on archive HarvestedDataArchive>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320145> <Size based data retirement operation completed on archive HarvestedDataArchive. Retired 0 records in 0 ms.>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320144> <Size based data retirement operation started on archive EventsDataArchive>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320145> <Size based data retirement operation completed on archive EventsDataArchive. Retired 0 records in 0 ms.>+
    +####<Feb 29, 2012 1:10:23 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <weblogic.cluster.MessageReceiver> <<WLS Kernel>> <> <> <1330539023098> <BEA-003107> <Lost 2 unicast message(s).>+
    +####<Feb 29, 2012 1:10:36 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330539036105> <BEA-000111> <Adding Pinellas1tMS2 with ID -9071779833610528123S:entwl2t-vm:[7005,7005,-1,-1,-1,-1,-1]:entwl2t-vm:7005,entwl3t-vm:7007:Pinellas1tDomain:Pinellas1tMS2 to cluster: Pinellas1tCluster1 view.>+
    +####<Feb 29, 2012 1:11:24 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[STANDBY] ExecuteThread: '3' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330539084375> <BEA-000128> <Updating -9071779833610528123S:entwl2t-vm:[7005,7005,-1,-1,-1,-1,-1]:entwl2t-vm:7005,entwl3t-vm:7007:Pinellas1tDomain:Pinellas1tMS2 in the cluster.>+
    +####<Feb 29, 2012 1:11:24 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[STANDBY] ExecuteThread: '4' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330539084507> <BEA-000128> <Updating -9071779833610528123S:entwl2t-vm:[7005,7005,-1,-1,-1,-1,-1]:entwl2t-vm:7005,entwl3t-vm:7007:Pinellas1tDomain:Pinellas1tMS2 in the cluster.>+
    Edited by: john wang on Feb 29, 2012 10:36 AM
    Edited by: john wang on Feb 29, 2012 10:37 AM
    Edited by: john wang on Feb 29, 2012 10:38 AM

    Hi John,
    There may be some circumstances like when there are many files in the WEB-INF folder and JPS don't use TLD.
    I don't think a 1hour deployment is normal, it should be much more faster.
    Since you are using 10.3.5, I suggesto you to install the corresponding patch:
    1. Download patch 10118941p10118941_1035_Generic.zip
    2. Uncompress the file p10118941_1035_Generic.zip
    3. Copy the required files (patch-catalog_XXXXX.xml, CIRF.jar ) to the Patch Download Directory (typically, this folder is <WEBLOGIC_HOME>/utils/bsu/cache_dir).
    4. Rename the file patch-catalog_XXXXX.xml into patch-catalog.xml .
    5. Start Smart Update from <WEBLOGIC_HOME>/utils/bsu/bsu.sh .
    6. Select "Work Offline" mode.
    7. Go to File->Preferences, and select "Patch Download Directory".
    8. Click "Manage Patches" on the right panel.
    9. You will see the patch in the panel below (Downloaded Patches)
    10. Click "Apply button" of the downloaded patch to apply it to the target installation and follow the instructions on the screen.
    11. Add "-Dweblogic.jsp.ignoreTLDsProcessingInWebApp=true" to the Java options to ignore additional findTLDs cost.
    12. Restart servers.
    Hope this helps.
    Thanks,
    Cris

  • File name too long cannot copy (cont'd)

    This is a continuation of the post started September 01, 2009, with the last post on
    October 17, 2011 named File name too long cannot copy
    Since this is an ongoing unsolved issue, and the thread was locked, it continues here.
    It is ever so easy to create a long file/path using explorer.exe as will be shown below.
    What needs to be solved is, what is an easy way (no, not by listing out all the files in a directory to a text file and counting characters), perhaps using a shell extension, to find out which files in a directory and it's subdirectories will be (or were)
    too long to copy, (and then perhaps even copying those over either intact or renamed)?
    Maflagulator said:
    I'm running the 7100 build...enjoying it except for one big thing:
    While attempting to copy 402gb from my main storage volume onto a spare 500gb drive (for the purpose of changing to a new RAID array) I've come across something that I would expect a Windows 98 OS to give me.
    It tells me that a file has TOO LONG of a file name, then provides with two unhelpful options: SKIP or CANCEL
    I never had XP give me an issue like this at all, so what gives? And while some specific files did have long file names (such as for songs, etc.) it had 7 issues with folders stating that their name was too long, but in fact they were not since they were
    titled '07-06-07' for the date that I dumped the audio files in them. However, they may have contained FILES with long file names though.
    Anyone else get this same situation? Perhaps the RTM version does not do this? Can anyone verify this regarding their install of the RC or the RTM?
    It made it through 400gb out of the 402gb transfer.
    I'm just happy to see that it doesn't spazz out about an issue like this until it has done all the other transfers that it can do because it saves the issues it has with files until the very end. In XP it would spazz about it the moment it came across it
    causing the transfer process to halt.
    Since long path/file names can so easily be created on Win7, it might be useful to see a typical way this happens, which might then give clues how to work with them.
    In Windows Vista, we learnt from:
    File names and file name extensions: frequently asked questions that:
    Windows usually limits file names to 260 characters. But the file name must actually be shorter than that, since the complete path (such as C:\Program Files\filename.txt) is included in this character count.
    In Windows 7, we are told here:
    File names and file name extensions: frequently asked questions that:
    It depends on the length of the complete path to the file (such as C:\Program Files\filename.txt). Windows limits a single path to 260 characters. This is why you might occasionally get an error when copying a file with a very long file name to a location
    that has a longer path than the file's original location.
    From the Windows Dev Center - Desktop, we read about Maximum Path Length Limitation here:
    Naming Files, Paths, and Namespaces
    This helps us understand why a folder can be a maximum of 244 characters, from the defined 260 length of MAX_PATH as follows:
    260 minus C:\ (3) minus <NUL> (1) = 256
    256 minus 8.3 file name (12) = 244
    We also learn there that: The Windows API has many functions that also have Unicode versions to permit an extended-length path for a maximum total path length of 32,767 characters.
    And we read the claim that: The shell and the file system have different requirements. It is possible to create a path with the Windows API that the shell user interface is not be able to interpret properly.
    There is also a comment below this document that reads: In a previous iteration of this document, it is mentioned that The Unicode versions of several functions permit a maximum path length of approximately 32,000 characters composed of components up to
    255 characters in length. This information is now gone.
    So we are in a position where the file system and Windows API can create long path/flies that the shell cannot handle.
    But then we need to be able to handle it, so a little exploration might lead to a better understanding of how to do this.
    For most tasks being performed on long folder/files, Windows 7 and other Windows programs balk when the Path+Filename length > 260
    Let's create a long path/file.
    Create a folder called A at the root of a Drive.
    Create a sub-folder of A called: B
    Create a sub-folder of B called: C
    Make a FILE in sub-folder C called (no spaces or break, one long continuous string): 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T
    123456789U123456789V123456789W123456789X123456.txt
    Rename sub-folder C to the string (no spaces or break, one long continuous string) (The actual directory created will be slightly shorter than this full length): 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J
    123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X123456789Y123456789Z
    Rename sub-folder B to the same full string above. (The actual directory created will be slightly shorter than this full length but 2 characters longer than the step above.)
    Rename folder A to that same full original string. (Again the actual directory created will be slightly shorter than this full length but 2 characters longer than the step above.)
    You now have the lovely file placed at (the breaks are just so it fits into the screen):
    C:\123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X1234\ 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J
    123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X12\ 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T
    123456789U123456789V123456789W123456789X\ 123456789A123456789B123456789C123456789D123456789E123456789F123456789G123456789H123456789I123456789J 123456789K123456789L123456789M123456789N123456789O123456789P123456789Q123456789R123456789S123456789T 123456789U123456789V123456789W123456789X123456.txt
    You have a folder length of over 700 and a file length of over 250 for a total of over 950
    However you will notice that each folder, when created, could only be a maximum of 247 charachters including the path (example C:\ , & C:\A , & C:\A\B
    This only applies backwards, that is up the path. It did not matter what was further down the path.
    Now, you can't easily access or rename the file, but you can rename the folders easily.
    For best results, start renaming from the top of the Tree, working down the subfolders, because renaming from down up will limit you, and in fact won't work if the folder lengths are too long.
    So how might knowing this help us?
    Well, to copy this long_file from the C:\ drive to the D:\ drive, and keeping the path structure, this should work:
    Note the name of the top folder. Rename it to something very short, say: A (Make sure C:\A does not exist)
    Note the name of the 2nd folder. Rename it to something very short, say: B (Make sure C:\A\B does not exist)
    Note the name of the 3rd folder. Rename it to something very short, say: C (Make sure C:\A\B\C does not exist)
    Make sure D:\A does not exist - then copy the A folder on disk C: to disk D: (which gives you D:\A\B\C\long_file
    Rename D:\A\B\C to D:\A\B\Original_3rd_Folder_name
    Rename D:\A\B to D:\A\B\Original_2nd_Folder_name
    Rename D:\A to D:\Original_top_Folder_name
    Rename C:\A\B\C back to their original names, in this same reverse order starting with C, then B, then A
    Note: If using Explorer, at some points you might have to press the F5 refresh key.
    This is of course how you might copy such long path/files without using the other more "easy" techniques for the "normal" everyday user like:
    sharing a sub-folder
    using the commandline to assign a drive letter by means of SUBST
    using AddConnectionunder VB to assign a drive letter to a path
    using the "\\?\" prefix to a path string to tell the Windows APIs to disable all string parsing and to send the string that follows it straight to the file system
    and so on.
    See how simple Windows can be for Dummies!
    But then, how can we know that files to be copied exceed this MAX_PATH? Or also after a copy has taken place, know exactly which files that have NOT been copied because of exceeding the MAX_PATH limit, and then a procedure to copy these either by renaming
    them, or by copying them intact as they are?
    There have been suggestions to use
    LongPathTool, but this does not have a facility to check a series of folders and tell you which files are going to be caught by the error when copying. So once a copy has taken place using Windows 7, one does not know which files did not get copied, and
    where exactly they are located.
    Neither does the free
    Old Path Scanner do that. It can only check for overly long directory paths, but misses out when the directory path is within limits, but adding in the file name puts it out of bounds.
    So, as shown above, it is ever so easy to create a long file/path using explorer.exe
    So, what then is an easy way (no, not by listing out all the files in a directory to a text file and counting characters), perhaps using a shell extension, to find out which files in a directory and it's subdirectories will be (or were) too long to copy,
    (and then perhaps even copying those over either intact or renamed)?

    This is a not a "solution" ....but a "work around": a low tech fix....for error message "your file name is too long to be copied, deleted, renamed, moved" :
    1.   problem is this: the "file name" has a limit on number of characters.....the sum of characters really includes the entire path name; you gotta shorten it first (i.e, if the total number of characters in the file name + Path name are over the
    limit, the error appears).  The deeper your file folder sub levels are, the more this problem will come up, especially when you copy to a subfolder of a subfolder/subfolder of another path ...adds to character limit)
    2.  How do you know which combined file names + path names are too long if  you are in the  middle of a copy operation and this error pops up?  Some files copied but the "long files error message" says "skip" or "cancel" ... but not which
    files are the "too long" ones.  If you hit "skip" or "cancel" the "too long" files are left behind...but are mixed in with non-offender "good" "short name" files.   Sorting thru 1000s of "good" files to find a few "bad" ones manually is impractical.
    3.   Here's how you sort out the "bad" from the "good":
    4.    Let's say you want to copy a folder ..."Football" ...that has five layers of subfolders; each subfolder contains numerous files:
      C:/1 Football / 2 teams/ 3 players/ 4 stats/ 5 injuriessidelineplayerstoolong 
           There are five levels root "1 football" with subfolders  2, 3, 4 and lastly "5 injuries"
    5.    Use "cut" and "paste"  (for example to backup all five levels to a new backup folder):
           select "1 football" ....cut....select target folder....paste 
           ("cut" command means as the files are copied to the target destination, the file is deleted from the source location)
          Hint: avoid "cut" and "paste" to a target folder that is itself a sub/sub/sub folder ...that compounds the "characters over the limit" problem ...because the characters in the sub/sub/sub folder are included in the "file name
    character limit"...instead "paste" to a C:/ root directory.
           Suppose in the middle of this operation...error pops up: "5 files have file names that are too long"  Skip or cancel?
           select "skip"  ...and let operation finish
    6.    Now go back and look at the source location: since the software allows only the "good" "short name" files to be copied (and because you "skipped" the "bad" "Long name" files so they are not copied or deleted) ...all that remains
    in the source location are the "bad" "long name files" (because "good" ones were deleted from the source location after the "cut" operation...the bad ones stick out like a sore thumb.
    7.   You will find ....all that remains in source folders are: the "bad" "too long" files; in this example the "bad" file is in level 5:
          C:/ 1 football / 2 teams /3 players /4 stats /5 injuriessidelineplayerstoolong
    8.   select folder 5 injuriessidelineplayerstoolong (that's right...select folder, not file) gotta rename the folder first.
    9.  hit F2 rename folder..folder name highlighted...delete some of the letters in the folder name:
           like this:   5 injuriessidelineplayers  ....you should delete 'toolong'....from the folder name
    10.  then go into folder 5....and do the same operation ...with the too long file name:
            hit F2 rename file....file name hightlighted...delete some of the letters
               like this:  injuriessidelineplayers.....you should delete 'toolong' from the file name
    11.  Now..."cut and paste"  the renamed file to the target backup folder.  
    The Error message will pop up again if you missed any "bad" files....for example, it will indicate "5 files too long" ....then repeat process 5 times until you fix all of them
    12.     Finally, copy the target destination files back to the source location (when you are certain all source location file folder locations are empty) 
    Of course, this "makeshift" solution would not be necessary if MSFT would fix the problem...

  • How can I download a 18 minute video I recorded from my iPad to either email or youtube or vimeo I keep getting the message the file is too long.

    How can I transfer a 18 minute video I recorded on my iPad camera app to email/youtube or vimeo? I keep getting the message the file is too long do want to send clips?
    Thanks

    Youtube usually had an 8 minute length cap, which may be part of your issue. Also at 18 minutes that is probably a gig+ file size, which will also take a long time to upload and may be another part of the issue.
    Edit it down as jim suggested. Or you may need to use something like dropbox, to get it off your ipad and onto a computer where might have more options to edit it down or size it down.
    If it's just file size you can transcode it into a smaller file - a lower quality MP4 for example. If it's the length that's an issue for youtube, you'll need to break it up into a couple different files and upload part 1 and part 2

  • File path too long when collecting??...

    Hi all,
    REALLY hope someone has some wisdom on this! I've searched forums and the web and not yet found an answer....
    I've got a MASSIVE project on an external drive that's on the way out, so I want to collect the project that's on it to another new healthy drive. The current file structure leaves alot to be desired (several years of updating the same project and splitting it into managable portions then importing them back together several times, files still on there that aren't in use in AE anymore but have been kept 'incase we need them again later', etc...you can imagine the mess!)
    SO, I'm getting the 'file path too long' error when trying to collect it. Now, I could manually go through every folder and sub folder and tidy things up within AE - this would days about a week I reckon! - but I'll do it if that's the only way. What I'm wondering is this: -
    A) is the problem AE is having with the deep structure *WITHIN* ae project, or with the actual folders/files on the disk itself?
    B) if I tidy it up in AE only will that fix it?
    C) is there an easier way to do this rather than manually going through everything and looking for the offenders???
    Thanks guys,
    Jim.

    GULP....so I have to reorganise my assets OUTSIDE of AE and then relink it all WITHIN AE afterwards?? With 3 x 430GB projects and thousands of assets that is going to take a looooooooonng time.
    What I don't understand (or like) is how AE will let you get this far without any warning and still operate/save, etc but then when you come to collect is goes 'urr urrrr, computer says no'.
    Is there really no way of automating some of this then? I mean even a script that highlighted items in the project that are over the maximum path length would save a hell of a lot of manual searching and counting.
    Thanks for your replies guys but.....AAAARRGGHH!!!!!

  • RTF export takes too long

    Hi all,
    you know why a report with a lot of data and many pages (about 300 pages), if I export this in PDF format using short time (1 min) and if I export this in PDF format takes too long (even 10 minutes) and sometimes even times out.
    Have an idea?
    Thanks

    re: Paul.  Not true, or shouldn't be.  XDCAM HD timeline, XDCAM HD output.
    re: Michael.  Well, close.  I mixdown the edited timeline, then place that into another sequence where I apply a Broadcast Safe filter and hit it with a bit of audio compression.  I render that, ridding myself of the dreaded "red lines".  However, when I splat-E, I get red lines all over again while I'm exporting.  When the export is done, the red lines disappear.  Now as an aside, I do experience actual "missing" render files, but that occurs when I quit a project and then re-open it later...they sometimes don't seem to re-link.  As I mentioned in this post, I have a feeling that's because this system is set up in the worst way possible (a single RAID5 which acts as system drive and media storage).
    Don't blame me, I'm just trying to work within the confines of what I'm given. My major concern here is that it seems to be getting worse.

  • OIM11g managed server startup takes too long

    I am running a OIM11g managed server using weblogic on windows xp; in development mode
    My managed server takes too long to startup; averaging atleast an hour at the minimum.
    I am attaching the logs from the server between the time frame when it took the longest time.
    Can someone please help me to improve the startup.
    Edited by: 838243 on Feb 20, 2011 10:22 AM

    <9:32:14 PM> <Info> <EJB><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-010008> <EJB Deploying file: iam-ejb.jar>
    <9:45:06 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '23' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <9:45:07 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:00:06 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '14' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <10:00:07 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:15:11 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '14' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <10:15:41 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:30:11 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '19' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <10:30:15 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:32:44 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149060> <Module iam-ejb.jar of application oim [Version=11.1.1.3.0] successfully transitioned from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:32:44 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149059> <Module dataobjects-ejb.jar of application oim [Version=11.1.1.3.0] is transitioning from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:33:08 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'dataobjects-ejb.jar' with uri, 'META-INF/weblogic-ejb-jar.xml'. The plan was merged.>
    <10:33:14 PM> <Info> <EJB><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-010008> <EJB Deploying file: dataobjects-ejb.jar>
    <10:38:10 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149060> <Module dataobjects-ejb.jar of application oim [Version=11.1.1.3.0] successfully transitioned from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:10 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149059> <Module iam-async-mdb.jar of application oim [Version=11.1.1.3.0] is transitioning from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:10 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'iam-async-mdb.jar' with uri, 'META-INF/ejb-jar.xml'. The plan was merged.>
    <10:38:11 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'iam-async-mdb.jar' with uri, 'META-INF/weblogic-ejb-jar.xml'. The plan was merged.>
    <10:38:11 PM> <Info> <EJB><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-010008> <EJB Deploying file: iam-async-mdb.jar>
    <10:38:13 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149060> <Module iam-async-mdb.jar of application oim [Version=11.1.1.3.0] successfully transitioned from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:13 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149059> <Module xlWebApp of application oim [Version=11.1.1.3.0] is transitioning from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:13 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'xlWebApp.war' with uri, 'WEB-INF/web.xml'. The plan was merged.>
    Edited by: 838243 on Feb 20, 2011 10:22 AM

  • Disk Utility (10.5.6) error: File name too long

    Hello, here is one for you all...
    I am in the midst of creating a new "build image" so I can start deploying Leopard.
    Before I upgraded the build to 10.5.6 I was able to create a disk image of this build machine.
    After updating it to 10.5.6 and all of the current Apple updates I am no long able to create a disk image.
    I keep getting an error several minutes into the imaging process that says +"*File name too long".*+
    I've run permission repair on the disk and Disk Warrior with still no luck. I've also used different machines and a boot FireWire drive to make he image and still have had no luck. I can use CCC to clone my Master build to another machine but I need to be able to make an image of it.
    Thanks in advance for any help.

    Bear in mind that in Unix, a file name includes the path, so if you have a very deep file or perhaps some kind of infinite recursive link, that might cause the problem, since that would make the file name very long indeed.
    This may not be the problem you are having, but it just came to my mind when I read your post

Maybe you are looking for

  • I can't get iTunes to work

    I got one of the new iPod Nanos for Christmas and got everything working(including iTunes). I put a bunch of music on didn't run iTunes very much for a while. The time came to put more music on, so I tried to open up iTunes, and it didn't come onscre

  • Shortcuts act differently

    Hi, I'm using a shortcut in one help file (File X.chm) to launch an external CHM file (FileY.chm). If you open the FileX.chm directly from Windows Explorer, the shortcuts work fine. But if you open the FileX.chm from within its related application, F

  • Publish to other than .Mac

    Is It possible to publish to another website other than .Mac like if you have your own website?????

  • Iphone 5 sprint unlock international

    hi i have sprint iphone 5 locked sprint carrier i contact with sprint support to unlock but they told me i have to sign in with your phone number i dont have the the number with my iphone i did'nt buy my iphone from sprint in my country we use locked

  • While trying to burn a cd in itunes, I receive the message "unknown error occurred" (-3)

    I had no trouble burning with Itunes until two days ago.  When I tried to burn, the disc was prepared, then the burn was cancelled with the message  "unknown error occurred" (-3) I use a speed of 2x to burn discs.