File Compression takes too long

Hi all,
I have a folder containing some matlab programs (around 200 of them), the file size is 4.6 MB. When I try to compress this folder (ctrl - click and use the default compression software), the compression never ends. I logged out of my account and re-logged in to see if it will help. But, it did not help. It happens only with this folder.
Any thoughts are appreciated.
Regards,
Murali

What happens if you take half of them and put them in a separate folder? Can you then zip either or both folders? I'm wondering if there is a bad file in the batch that is causing the compressor to choke.
Francine
Francine
Schwieder

Similar Messages

  • File.exists() takes too long to return.

    Hi.
    I have written a program which checks to see if a particular shared file is present on a number of different computers. The path to the file is the same on all the computers.
    I construct the File with IP + path.
    Then I call file.exists()
    If the remote computer is on and the file is present, then 'true' is returned quite quickly. But if the file is not available then 'false' is returned but only after about 30 seconds. This is too long and I want to set some sort of time out for the method after about 3 seconds. So that if the file is not found before the timeOut time, 'false' will be returned.
    Is there any way I can do this???
    Thanks,
    Jim

    Hi,
    I am already using Threads, here is my code:
    public class FileExistsThread extends Thread {
         private File file;
         private long tryTime;
         private boolean exists;
         private boolean changed;
         public FileExistsThread(File file, long tryTime) {
              this.file = file;
              this.tryTime = tryTime;
         public void run() {
              new Thread() {
                   public void run() {
                        try {
                             Thread.sleep(tryTime);
                        } catch(Exception e) {}
                        exists = false;
                        changed = true;
              }.start();
              exists = file.exists();
              changed = true;
         public boolean fileExists() {
              start();
              while(!changed);
              return exists;
    }The Problem is that I cant close my program while a file.exists() method is still executing no matter what Thread it is in. Say I'm looking for a file which the program will take 30 seconds to establish is not available. Then for that 30 seconds my program will refuse to close. Sometimes it can be much longer than this.
    This is why I still need some sort of timeout mechanism.
    Any Ideas?
    Cheers,
    Jim

  • I don't know why it takes too long time to sample flat file.

    I don't know why it takes too long time to sample flat file.
    OWB Client 10.1
    While importing a flat file of fixed width ,
    in the screen "Flat File Sample Wizard" shows the text box number of rows with default value 200.
    I want to extend this value to 700,000.
    But, it takes too long time (over 5 hours) to sample it.
    Do you know why it is happend? or How can i fix this problem?
    Thanks in Advance.
    Regards,
    JWS.

    Hello,
    Actually flat file sampling process’ goal is to capture the structure of the file. That’s why initially the sample size is set to 200 lines.
    The question is why you are trying to perform sampling by 700000 rows? Are you expecting some change in structure beyond this mark?
    If so, and you want to capture the fact that your source file is multi – typed, your better prepare small file for sampling outside the OWB.
    Sergey

  • HT1386 copying files in itunes takes too long

    When I drag music files into the iPhone's playlist, the copying takes too long.  I am running 11 and have a 4

    Wait on the update to finish downloading and installing.
    If the device is in recovery mode there is really nothing that can be done besides restoring it.

  • File Path browse take too long time

    Hi all,
    I am new one and I just start  learning and implementing labview program. Now I am develping data convertion program. My program purpose is 
    (1) reading data from input text file
    (2) do data calculation and convertion as the required format.
    (3) formatted data write to text file or Excel file.
    (a) I use' file path  ' from control palette. When I click on " browse folder" icon to choose the path of my input file or output file,it took too long to get there. I already tested and tested again . but I can't find the solution.
     (b) when I write calculated data to text file or excel, for a big amount of data, text file is ok. excel sheet show error if data is over 65536 rows. I want to know how to create and add data to new excel sheet if current sheet is full
    So, I attach my project and sample input text file. I really need help from you all.  Thanks
    Attachments:
    01_02_2010_Extract_Data1.vi ‏85 KB
    test3.txt ‏7 KB

    Your code needs a good cleanup.  I suspect the reason that your File Path browse button doesn't respond quickly is because you don't have any wait inside your while loop.  That means the loop is running as fast as it possibly can, even though you don't even have a file open to process, which keeps the processor busy and prevents it from doing anything else (like opening the file selection dialog).  Try adding a 100ms wait somewhere in the big while loop.
    It's good practice to keep your LabVIEW diagrams smaller than the size of your screen.  You have a lot of bends in your wires, which makes them hard to follow.  You don't need an enormous sequence structure outside of everything since it only has one frame, and in fact you probably don't need any sequence structures at all.  In your for loop, use a shift register instead of reading from and writing to a local variable.   Use the "Spreadsheet String to Array" function to separate your tab-delimited data; it will be much easier than the way you're doing it.
    The 65536 row limit is in Excel.  The easiest thing for you to do is to keep track of the number of rows you have written to the file, and whenever it reaches 65536, create a new file.  Since you're writing to a text file there's no way to create a new sheet.  To create a real Excel workbook with multiple sheets, you will need to look at the examples of how to use ActiveX to control Excel.

  • RTF export takes too long

    Hi all,
    you know why a report with a lot of data and many pages (about 300 pages), if I export this in PDF format using short time (1 min) and if I export this in PDF format takes too long (even 10 minutes) and sometimes even times out.
    Have an idea?
    Thanks

    re: Paul.  Not true, or shouldn't be.  XDCAM HD timeline, XDCAM HD output.
    re: Michael.  Well, close.  I mixdown the edited timeline, then place that into another sequence where I apply a Broadcast Safe filter and hit it with a bit of audio compression.  I render that, ridding myself of the dreaded "red lines".  However, when I splat-E, I get red lines all over again while I'm exporting.  When the export is done, the red lines disappear.  Now as an aside, I do experience actual "missing" render files, but that occurs when I quit a project and then re-open it later...they sometimes don't seem to re-link.  As I mentioned in this post, I have a feeling that's because this system is set up in the worst way possible (a single RAID5 which acts as system drive and media storage).
    Don't blame me, I'm just trying to work within the confines of what I'm given. My major concern here is that it seems to be getting worse.

  • Web application deployment takes too long?

    Hi All,
    We have a wls 10.3.5 clustering environment with one admin server and two managered servers separately. When we try to deploy a sizable web application, it takes about 1 hour to finish. It seems that it takes too long to finish the deployment. Here is the output from one of two managerd server system log. Could anyone tell me it is normal or not? If not, how can I improve this?
    Thanks in advance,
    John
    +####<Feb 29, 2012 12:11:03 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535463373> <BEA-149059> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] is transitioning from STATE_NEW to STATE_PREPARED on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:05 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <9baa7a67b5727417:26f76f6c:135ca05cff2:-8000-00000000000000b0> <1330535465664> <BEA-149060> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] successfully transitioned from STATE_NEW to STATE_PREPARED on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466493> <BEA-149059> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] is transitioning from STATE_PREPARED to STATE_ADMIN on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466493> <BEA-149060> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] successfully transitioned from STATE_PREPARED to STATE_ADMIN on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466809> <BEA-149059> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] is transitioning from STATE_ADMIN to STATE_ACTIVE on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 12:11:06 PM EST> <Info> <Deployer> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330535466809> <BEA-149060> <Module copyrequest of application copyrequest [Version=COPYREQUEST0002bb] successfully transitioned from STATE_ADMIN to STATE_ACTIVE on server Pinellas1tMS3.>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442300> <BEA-320143> <Scheduled 1 data retirement tasks as per configuration.>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320144> <Size based data retirement operation started on archive HarvestedDataArchive>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320145> <Size based data retirement operation completed on archive HarvestedDataArchive. Retired 0 records in 0 ms.>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320144> <Size based data retirement operation started on archive EventsDataArchive>+
    +####<Feb 29, 2012 1:00:42 PM EST> <Info> <Diagnostics> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '1' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330538442301> <BEA-320145> <Size based data retirement operation completed on archive EventsDataArchive. Retired 0 records in 0 ms.>+
    +####<Feb 29, 2012 1:10:23 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <weblogic.cluster.MessageReceiver> <<WLS Kernel>> <> <> <1330539023098> <BEA-003107> <Lost 2 unicast message(s).>+
    +####<Feb 29, 2012 1:10:36 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[ACTIVE] ExecuteThread: '2' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330539036105> <BEA-000111> <Adding Pinellas1tMS2 with ID -9071779833610528123S:entwl2t-vm:[7005,7005,-1,-1,-1,-1,-1]:entwl2t-vm:7005,entwl3t-vm:7007:Pinellas1tDomain:Pinellas1tMS2 to cluster: Pinellas1tCluster1 view.>+
    +####<Feb 29, 2012 1:11:24 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[STANDBY] ExecuteThread: '3' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330539084375> <BEA-000128> <Updating -9071779833610528123S:entwl2t-vm:[7005,7005,-1,-1,-1,-1,-1]:entwl2t-vm:7005,entwl3t-vm:7007:Pinellas1tDomain:Pinellas1tMS2 in the cluster.>+
    +####<Feb 29, 2012 1:11:24 PM EST> <Info> <Cluster> <entwl3t-vm.co.pinellas.fl.us> <Pinellas1tMS3> <[STANDBY] ExecuteThread: '4' for queue: 'weblogic.kernel.Default (self-tuning)'> <<WLS Kernel>> <> <> <1330539084507> <BEA-000128> <Updating -9071779833610528123S:entwl2t-vm:[7005,7005,-1,-1,-1,-1,-1]:entwl2t-vm:7005,entwl3t-vm:7007:Pinellas1tDomain:Pinellas1tMS2 in the cluster.>+
    Edited by: john wang on Feb 29, 2012 10:36 AM
    Edited by: john wang on Feb 29, 2012 10:37 AM
    Edited by: john wang on Feb 29, 2012 10:38 AM

    Hi John,
    There may be some circumstances like when there are many files in the WEB-INF folder and JPS don't use TLD.
    I don't think a 1hour deployment is normal, it should be much more faster.
    Since you are using 10.3.5, I suggesto you to install the corresponding patch:
    1. Download patch 10118941p10118941_1035_Generic.zip
    2. Uncompress the file p10118941_1035_Generic.zip
    3. Copy the required files (patch-catalog_XXXXX.xml, CIRF.jar ) to the Patch Download Directory (typically, this folder is <WEBLOGIC_HOME>/utils/bsu/cache_dir).
    4. Rename the file patch-catalog_XXXXX.xml into patch-catalog.xml .
    5. Start Smart Update from <WEBLOGIC_HOME>/utils/bsu/bsu.sh .
    6. Select "Work Offline" mode.
    7. Go to File->Preferences, and select "Patch Download Directory".
    8. Click "Manage Patches" on the right panel.
    9. You will see the patch in the panel below (Downloaded Patches)
    10. Click "Apply button" of the downloaded patch to apply it to the target installation and follow the instructions on the screen.
    11. Add "-Dweblogic.jsp.ignoreTLDsProcessingInWebApp=true" to the Java options to ignore additional findTLDs cost.
    12. Restart servers.
    Hope this helps.
    Thanks,
    Cris

  • OIM11g managed server startup takes too long

    I am running a OIM11g managed server using weblogic on windows xp; in development mode
    My managed server takes too long to startup; averaging atleast an hour at the minimum.
    I am attaching the logs from the server between the time frame when it took the longest time.
    Can someone please help me to improve the startup.
    Edited by: 838243 on Feb 20, 2011 10:22 AM

    <9:32:14 PM> <Info> <EJB><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-010008> <EJB Deploying file: iam-ejb.jar>
    <9:45:06 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '23' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <9:45:07 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:00:06 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '14' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <10:00:07 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:15:11 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '14' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <10:15:41 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:30:11 PM> <Info> <JDBC><oim1> <[ACTIVE] ExecuteThread: '19' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-001128> <Connection for pool "mds-oim" closed.>
    <10:30:15 PM> <Info> <Common><oim1> <MDSPollingThread-[oim, jdbc/mds/MDS_REPOS]> <anonymous> <BEA-000628> <Created "1" resources for pool "mds-oim", out of which "1" are available and "0" are unavailable.>
    <10:32:44 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149060> <Module iam-ejb.jar of application oim [Version=11.1.1.3.0] successfully transitioned from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:32:44 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149059> <Module dataobjects-ejb.jar of application oim [Version=11.1.1.3.0] is transitioning from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:33:08 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'dataobjects-ejb.jar' with uri, 'META-INF/weblogic-ejb-jar.xml'. The plan was merged.>
    <10:33:14 PM> <Info> <EJB><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-010008> <EJB Deploying file: dataobjects-ejb.jar>
    <10:38:10 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149060> <Module dataobjects-ejb.jar of application oim [Version=11.1.1.3.0] successfully transitioned from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:10 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149059> <Module iam-async-mdb.jar of application oim [Version=11.1.1.3.0] is transitioning from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:10 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'iam-async-mdb.jar' with uri, 'META-INF/ejb-jar.xml'. The plan was merged.>
    <10:38:11 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'iam-async-mdb.jar' with uri, 'META-INF/weblogic-ejb-jar.xml'. The plan was merged.>
    <10:38:11 PM> <Info> <EJB><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-010008> <EJB Deploying file: iam-async-mdb.jar>
    <10:38:13 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149060> <Module iam-async-mdb.jar of application oim [Version=11.1.1.3.0] successfully transitioned from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:13 PM> <Info> <Deployer><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-149059> <Module xlWebApp of application oim [Version=11.1.1.3.0] is transitioning from STATE_NEW to STATE_PREPARED on server oim_server1.>
    <10:38:13 PM> <Info> <J2EE><oim1> <[ACTIVE] ExecuteThread: '0' for queue: 'weblogic.kernel.Default (self-tuning)'> <WLS Kernel> <BEA-160166> <References were found in deployment plan for module 'xlWebApp.war' with uri, 'WEB-INF/web.xml'. The plan was merged.>
    Edited by: 838243 on Feb 20, 2011 10:22 AM

  • Why does it take too long to open attachments from email account?

    It takes forever to download attachments from my email account. I have tried to do the following tasks explained on the vista forum :http://windowshelp.microsoft.com/Windows/en-US/help/6b046ae9-1434-4423-9303-400ff6fe686b1033.mspx#ESD but none of the possible fixes work.
    After clicking open on the pop up box asking whether i want to open or save the attachement It takes too long to download it. The transfer window stays open showing that it is ready to download but stays at that bhox window. I press cancel and try to open again, if lucky it opens that file, otherwise it takes forever forceing me to cancel. The files very small files most of the time, usually between 50kb so should take seconds.
    I have even tried to save the files but again same process. The transfer box stays open but does not download.
    Any one any ideas.
    Thanx in advance.

    Hello
    It is not easy to say what happen exactly but it must be something with email account provider and their page. For me this case is not some typical Vista problem but you can try to find solution on Microsoft Vista IT Pro forum.
    By the way: do you have alternative mail address by some other provider? Is there the same situation?

  • Why finding replication stream matchpoint takes too long

    hi,
    I am using bdb je 5.0.58 HA(two nodes group,JVM 6G for each node).
    Sometimes, I found bdb node takes too long to restart(about 2 hours).
    When this occurs, I catch the process stack of bdb(jvm process) by jstack.
    After analyzing stack,I found "ReplicaFeederSyncup.findMatchpoint()" taking all the time.
    I want to know why this method takes so much time,and how can I avoid this bad case.
    Thanks.
    帖子经 liang_mic编辑过

    Liang,
    2 hours is indeed a huge amount of time for a node restart. It's hard be sure without doing more detailed analysis of your log as to what may be going wrong, but I do wonder if it is related to the problem you reported in outOfMemory error presents when cleaner occurs [#21786]. Perhaps the best approach is for me to describe in more detail what happens when a replicated node is connecting with a new master, which might give you more insight into what is happening in your case.
    The members of a BDB JE HA replication group share the same logical stream of replicated records, where each record is identified with a virtual log sequence number, or VLSN. In other words, the log record described by VLSN x on any node is the same data record, although it may be stored in a physically different place in the log of each node.
    When a replica in a group connects with a master, it must find a common point, the matchpoint, in that replication stream. There are different situations in which a replica may connect with a master. For example, it may have come up and just joined the group. Another case is when the replica is up already but a new master has been elected for the group. One way or another, the replica wants to find the most recent point in its log, which it has in common with the log of the master. Only certain kinds of log entries, tagged with timestamps, are eligible to be used for such a match, and usually, these are transaction commits and aborts.
    Now, in your previous forum posting, you reported an OOME because of a very large transaction, so this syncup issue at first seems like it might be related. Perhaps your replication nodes need to traverse a great many records, in an incomplete transaction, to find the match point. But the syncup code does not blindly traverse all records, it uses the vlsn index metadata to skip to the optimal locations. In this case, even if the last transaction was very long, and incomplete, it should know where the previous transaction end was, and find that location directly, without having to do a scan.
    As a possible related note, I did wonder if something was unusual about your vlsn index metadata. I did not explain this in outOfMemory error presents when cleaner occurs but I later calculated that the transaction which caused the OOME should only have contained 1500 records. I think that you said that you figured out that you were deleting about 15 million records, and you figured out that it was the vlsn index update transaction which was holding many locks. But because the vlsn index does not record every single record, it should only take about 1,500 metadata records in the vlsn index to cover 15 million application data records. It is still a bug in our code to update that many records in a single transaction, but the OOME was surprising, because 1,500 locks shouldn't be catastrophic.
    There are a number of ways to investigate this further.
    - You may want to try using a SyncupProgress listener described at http://docs.oracle.com/cd/E17277_02/html/java/com/sleepycat/je/rep/SyncupProgress.html to get more information on which part of the syncup process is taking a long time.
    - If that confirms that finding the matchpoint is the problem, we have an unadvertised utility, meant for debugging, to examine the vlsn index. The usage is as follows, and you would use the -dumpVLSN option, and run thsi on the replica node. But this would require our assistance to interpret the results. We would be looking for the records that mention where "sync" points are, and would correlate that to the replica's log, and that might give more information if this is indeed the problem, and why the vlsn index was not acting to optimize the search.
    $ java -jar build/lib/je.jar DbStreamVerify
    usage: java { com.sleepycat.je.rep.utilint.DbStreamVerify | -jar je-<version>.jar DbStreamVerify }
    -h <dir> # environment home directory
    -s <hex> # start file
    -e <hex> # end file
    -verifyStream # check that replication stream is ascending
    -dumpVLSN # scan log file for log entries that make up the VLSN index, don't run verify.
    -dumpRepGroup # scan log file for log entries that make up the rep group db, don't run verify.
    -i # show invisible. If true, print invisible entries when running verify mode.
    -v # verbose

  • HT1338 audio midi initializing takes too long

    I am having trouble with midisport 4x4 installation and setup. My Cubase 6 (and even garageband) take too long to initialize audio midi setup and finally don't succeed.As a result midisport 4x4 don't work. Even when it does there is notable latency.My audio interface is DIGIDESIGN 003 Rack. I've tried Avid MBOX PRO also. Problem remain the same.

    Hi Brad,
    Thank you for your quick response. I am sorry that I couldn't upload the file, and it also has some VI involved with as well, so I don't know if I can do it.
    Since it doesn't work when I add 2 inspections to one, I decide go with the Select Inspection Function now; and it seems working well and doesn't take much time to load each of the inspection.
    Thanks.

  • Sometimes my computer takes too long to connect to new website. I am running a pretty powerful work program at same time, what is the best solution? Upgrading speed from cable network, is it a hard drive issue? do I need to "clean out" the computer?

    Many times my computer takes too long to connect to new website. I have wireless internet (time capsule) and I am running a pretty powerful real time financial work program at same time, what is the best solution? Upgrading speed from cable network? is it a hard drive issue? do I only need to "clean out" the computer? Or all of the above...not to computer saavy.  It is a Macbook Pro  osx 10.6.8 (late 2010).

    Almost certainly none of the above!  Try each of the following in this order:
    Select 'Reset Safari' from the Safari menu.
    Close down Safari;  move <home>/Library/Caches/com.apple.Safari/Cache.db to the trash; restart Safari.
    Change the DNS servers in your network settings to use the OpenDNS servers: 208.67.222.222 and 208.67.220.220
    Turn off DNS pre-fetching by entering the following command in Terminal and restarting Safari:
              defaults write com.apple.safari WebKitDNSPrefetchingEnabled -boolean false

  • Accessing BKPF table takes too long

    Hi,
    Is there another way to have a faster and more optimized sql query that will access the table BKPF? Or other smaller tables that contain the same data?
    I'm using this:
       select bukrs gjahr belnr budat blart
       into corresponding fields of table i_bkpf
       from bkpf
       where bukrs eq pa_bukrs
       and gjahr eq pa_gjahr
       and blart in so_DocTypes
       and monat in so_monat.
    The report is taking too long and is eating up a lot of resources.
    Any helpful advice is highly appreciated. Thanks!

    Hi max,
    I also tried using BUDAT in the where clause of my sql statement, but even that takes too long.
        select bukrs gjahr belnr budat blart monat
         appending corresponding fields of table i_bkpf
         from bkpf
         where bukrs eq pa_bukrs
         and gjahr eq pa_gjahr
         and blart in so_DocTypes
         and budat in so_budat.
    I also tried accessing the table per day, but it didn't worked too...
       while so_budat-low le so_budat-high.
         select bukrs gjahr belnr budat blart monat
         appending corresponding fields of table i_bkpf
         from bkpf
         where bukrs eq pa_bukrs
         and gjahr eq pa_gjahr
         and blart in so_DocTypes
         and budat eq so_budat-low.
         so_budat-low = so_budat-low + 1.
       endwhile.
    I think our BKPF tables contains a very large set of data. Is there any other table besides BKPF where we could get all accounting document numbers in a given period?

  • Report Takes Too Long Time

    Hi!
    I am in troubble
    following is the query
    SELECT inv_no, inv_name, inv_desc, i.cat_id, cat_name, i.sub_cat_id,
    sub_cat_name, asset_cost, del_date, i.bl_id, gen_desc bl_desc, p.prvcode, prvdesc, cur_loc,
    pldesc, i.pmempno, pmname, i.empid, empname
    FROM inv_reg i,
    cat_reg c,
    sub_cat_reg s,
    gen_desc_reg g,
    ploc p,
    province r,
    pmaster m,
    iemp_reg e
    WHERE i.sub_cat_id = s.sub_cat_id
    AND i.cat_id = s.cat_id
    AND s.cat_id = c.cat_id
    AND i.bl_id = g.gen_id
    AND i.cur_loc = p.plcode
    AND p.prvcode = r.prvcode
    AND i.pmempno = m.pmempno(+)
    AND i.empid = e.empid(+)
    &wc
    order by prvdesc, pldesc, cat_name, sub_cat_name, inv_no
    and this query returns 32000 records
    when i run this query on reports 10g
    then it takes 10 to 20 minuts to generate report
    how can i optimize it...?

    Hi Waqas Attari
    Pls study & try this ....
    When your query takes too long ...
    hope it helps....
    Regards,
    Abdetu...

  • OPM process execution process parameters takes too long time to complete

    PROCESS_PARAMETERS are inserted every 15 min. using gme_api_pub packages. some times it takes too long time to complete the batch ,ie completion of request. it takes about 5-6 hrs long time ,in other time s it takes only 15-20 mins.This happens at regular interval...if anybody can guide me I will be thankful to him/her..
    thanks in advance.
    regds,
    Shailesh

    Generally the slowest part of the process is in the extraction itself...
    Check in your source system and see how long the processes are taking, if there are delays, locks or dumps in the database... If your source is R/3 or ECC transactions like SM37, SM21, ST22 can help monitor this activity...
    Consider running less processes in parallel if you have too many and see some delays in jobs... Also indexing some of the tables in the source system to expedite the extraction, make sure there are no heavy processes or interfaces running in the source system at the same time you're trying to load... Check with your Basis guys for activity peaks and plan accordingly...
    In BW also check in your SM21 for database errors or delays...
    Just some ideas...

Maybe you are looking for