Performance issues (excessively slow)

Acrobat Professional is not running anywhere near as smoothly as one might expect, and I'm looking for a reason why.
I'm using a Pentium 4 2.4Ghz work computer, on Windows XP Home, version 2002 Service Pack 3, with 1Gb of ram and 55Gb free space. I've installed Acrobat Pro version 9.1.1.
When simply editing existing .pdf files as part of a catalogue, all I'm doing is copy&paste, drag&drop, insert image, and edit text. Nothing more. But even so, I'm seeing wait times of up to 30 seconds between clicks. After simply pasting a template into the file and dragging it into place, Acrobat freezes up for up to 30 seconds before I'm able to do anything more. Zooming in, something that involves only holding down a key and scrolling a mouse wheel, takes a minimum of 20 seconds.
I've managed to convince my employers to purchase a more powerful tower, but they need some idea of just how powerful it needs to be. This one is already twice as fast as the recommended system requirements, yet I'm having such trouble.
Or is there a way to solve this problem without upgrading?

Personally, I think 1 GB for XP is a minimal system---I don't care what the minimal system requirements are. Additionally, you probably have programs in the background that are taking up precious resources (minimally, you should have a firewall, virus protection, spyware protection, and the acrotray and who knows what else). This all takes precious resources that Acrobat and XP needs.  Additionally, there is something very wrong if your workflow requires you to edit a pdf catlogue via copy & paste, drag and drop, and editing text. This is what most of us here consider extensive editing of a pdf file. All editing should whenever possible be done in the original program---NOT in Acrobat. Yes Acrobat provides the tools but they should be reserved for emergencies. Acrobat is not at DTP, graphics or WP program and should not be treated as such. i bet you will find going back to the original files will be much faster even if you though you need to regenerate the pdf files.

Similar Messages

  • Camileo X Sports - SD card performance issue - too slow

    Hello,
    can anybody help me please.....when i tun on the cam and make the FIRST record it always shows me that the speed card is too slow, but i have sandisk 64gb extreme with 48mbs.....but it only dissapears at the first record and not the following.
    what can i do....i have also the newest firmware.

    yup! I had the same problem at first, i have the Samy EVO class 10 32G SDHC.
    It's usually brilliant, actually it works fine with the 50 fps, 30 fps and even the 720p 120 fps.
    However i found the problem easier than most people simply due to sheer coincidence.
    I left the screen on while recording and right before i stopped recording (btw it never froze in my case it just stopped recording) and the screen displayed "low speed card"! what do you know!
    I researched the issue and it turns out the Samsun EVO was a bit slower than what the camera required at 60 fps. at 50 fps it barely stopped recording once.
    SanDisk Micro SDHC Extreme Class 3 32GB SDSDQXN-032G-G46A
    This card is a relatively cheap one, it is a UHS class 3 which mean it's an Ultra High Speed card. The class 3 is a great ranking. It's like 10 Euros more expensive than the Samsun and from what i've read works great with the GoPro hero 4 while shooting 4K & 1080p at 120 fps so it should hold up very well.
    It was never the upgrade or downgrade of the software. Lucky i left the screen on!

  • ITunes on Windows 7 performance issues - slow and hangs

    Is there any way to improve performance of iTunes running on Windows 7?
    It has always been slow with lots of application hangs and stalls that will resolve themselves if you wait for iTunes to catch up, but I find it incredibly frustrating.
    For example, when iTunes starts and if it is checking podcasts, the entire application is unresponsive until it finishes checking and refreshing podcasts.
    Alternately, when viewing my iPad Air and looking at the list of apps installed via iTunes it can take 30 seconds before iTunes responds.
    Considering that this is running on a system with a quad-core Haswell processor at 3.2Ghz, 16GB of RAM and Windows 7 x64 it is pretty unacceptable.
    The efficiency at which iTunes operates is just appalling.  I can run Adobe Photoshop and many more memory hungry and processor intensive applications than iTunes and get better performance than what I see in iTunes.
    This is not specific to my system either.  All of my family members have iPhones and their own computers and this performance issue exists on every computer I have ever run iTunes on.
    Is there any way to manually tune performance because this is pretty crappy.  It's been bugging me for years, but today just kind of feels like the last straw.

    I guess the answer is, "It's terribad, live with it"

  • While creating Billing, system is very slow..performance issue

    Hi,
    While creating Billing, system is very slow. How can I debugg and provide the analysis where is the exact problem.
    This is showing performance issue.
    Waiting for kind response.
    Best Regards,
    Padhy
    Moderator Message : Duplicate post locked.
    Edited by: Vinod Kumar on May 12, 2011 10:59 AM

    hi,
    Chk the links
    http://help.sap.com/saphelp_nw04/helpdata/en/4a/e71f39488fee0ce10000000a114084/content.htm
    Re: How to create Secondary Index?
    How may secondary indices I can create on the ODS?
    Deletion of ODS index
    Ramesh

  • Slow Moving Items in 0IC_MC01 Performance issue

    Hi  i am creating slow moving items report (Report which shows materials for which no transaction has been done for last 3 years) on MP 0IC_MC01 in BI7.0.
    MP is based on cube 0IC_C03 and Infoobject 0MATERIAL .
    In query i put rowcount in coloumn with Infoprovider 0material with constant selection but i dont get the option "Display values not posted" . and Issues value and Recved value .
    in rows i put plant , material and material type 
    n put one condition values for issued and Recid = 0 .
    and in filter we have calender year restriction  .
    But the proble is dere are around 25000 materials for each plant and its taking a lot of time to exceute for multiple plants for 1-2 years , can anyone suggest the best way for this report or how performance of this report can be improved .
    Help will be appreciated .

    Its rather better to use infoset on 0IC_MC01 and Material than using MP. The performance issue is coming because of a join between cube and infoobject, which SAP dont suggest to use.
    You ahve to use left outer join to get slow moving items.
    http://help.sap.com/saphelp_nw04/Helpdata/EN/67/7e4b3eaf72561ee10000000a114084/content.htm
    Thanks...
    Shambhu

  • Yet Another Slow Performance Issue

    I am noticing some seriously slow performance on my state-of-the-art 2.5 Ghz G5. I don't even know where to start. It happens frequently, it happens switching back and forth b/w apps, I get the spinning ball all the time, I have to restart to "refrech" the machine, which doesn't help, it hangs, it thinks, it does whatever it does. No real reason it would be doing this, I am not taxing the processors in any great way. I only have a few apps open, running a medium size photoshop file, and I'm trying to get some work done on a deadline and it's really reaaly irritating.
    What kind of maintenance should I run to cleanse this machine or somthing? Norton? De-fragment? I sure hope there's a solution for this weak performance.
    G5, dual 2.5Ghz   Mac OS X (10.4.2)  

    Norton Utilties hasn't been good to us, even their AV has had a serious takeover exploit because it runs with root powers.
    Applications running with root powers (admin password install) is one of the greatest threats we face as Mac users because it introduces all the more chances of failure behind our greatest security barrier, the admin password.
    About your performance issues, I have written something about it that can make your Mac fly like the wind.
    Visit my site and download the performance and cloning text doc, it's a bit of rough draft, but the hard core info is there.
    One thing I need to add to my file above is that Western Digital is coming out with a larger faster drive called the RaptorX. Even though the clear cover is about worthless to us PowerMac G5 owners, the size and speed of this drive makes a excellent boot drive for performance.

  • Service Manager 2012 SP1 consoles hanging or slow performance issue in Virtual Environment

    Hii,
    We are facing SCSM SP1 console performance issue which is utilizing high CPU and working deadly.
    For information, our SCSM is in Virtual environment via Hyper-V.
    When running the console over an RDP session to a Hyper-V virtual machine, we have to be careful not to maximize the console so that it will remain fast.  If we maximize on the VM, the console is so slow as for it to be unusable.
    Can someone share his experience, please?
    Regards, Syed Fahad Ali

    Hi Sayed,
    This is a bug and hopefully Microsoft team will solve it soon. if you can to vote for this bug here
    https://connect.microsoft.com/WindowsServer/feedback/details/810667/scsm-console-consumes-a-lot-of-cpu-when-opened-maximized-on-work-item-view-like-all-incidents
    Mohamed Fawzi | http://fawzi.wordpress.com

  • Unraveling slow RoboSource performance issues

    Greetings all,
    I am a Robo-noob (noobie, novice, etc) though not new to
    computers and databases and such. We have a team of 5 writers all
    working on separate projects but under the same RoboSource version
    control. We are experiencing some serious performance issues:
    projects can take over an hour to check out or in; not all users
    can see projects; "rename" doesn't get rid of the old rather makes
    a copy under the new name(?); etc.
    I am trying to baseline "normal" behavior for RoboSource
    Control as well as some best practices for multiple users. I am
    wondering if there is something more in depth to describe the
    RoboSource control engine, and possibly maintenance tasks to help
    what appears to be a jumble of nodes that are not all necessary,
    too many 'admin' users, etc. I expect that "cleanup" will solve a
    lot of our problems.
    So, rather than go into each problem in the forum, I would
    prefer to find good resource and tackle this thing iteratively.
    Thanks for any useful suggestions
    Don

    Hi Don
    From what you posted, it would seem to indicate each of the 5
    help authors are "captiain of their own ship" so to speak, and are
    only responsible for their individual projects. So why the use of
    RoboSource control? Why even introduce this layer of complexity in
    the mix? What does it offer you need?
    I really wish I had some nifty advice here about what to do.
    But I tend to only show folks how to access RSC in the classes I
    teach. To be very honest, if it were me, I'd probably opt for using
    something like Microsoft Visual Source Safe. And I actually have
    used VSS with RoboHelp in the past. Long before RoboHelp offered
    any "integration" or its own RSC.
    I'm not saying RSC is a "bad" product here. Only that there
    seem to be far more issues reported with it than there are answers.
    And performance is certainly among the cited issues. I'm sure you
    may well have your reasons for using it (or any other product for
    that matter) but based on what little we know at this point, it
    would seem overkill.
    Sincerely... Rick

  • Cache-flush VM-related performance issue

    Dear forum,
    I've got a peculiar performance issue going on with the BDB pagecache being flushed to disk. I've managed to reproduce the issue perfectly on three out of three quite different systems that I've tried on, so it is at least quite well-defined.
    My usage pattern for the database in question is such that I periodically (perhaps once every 10-60 seconds or so) need to read through an amount of values (around 500-2000 or so) from a database containing a rather large amount (in the millions, at least) of keys. There are a few writes for every such batch, but not very many (a couple of tens). The keys that are read each batch are quite random, and very likely to be completely different from batch to batch. The database is a DB_HASH.
    When I do that, BDB seems to dirty a lot of pages in the page cache (which I have currently sized at 512 MB so that pages don't have to be forced out from it), from what I can tell by manipulating refcounts and stuff, so all in all, a single batch seems to dirty some 10-40 MB or so of the mmapped cache region. (I check this using pmap -x on Linux.) Note that when I speak of pages and the dirtying of them here, I mean at the VM level, not the BDB level.
    A while after this has happened, the VM comes around and wants to flush the dirty pages to disk, so it batches writes of large portions (often the entire set of dirty pages, but sometimes it only does 10-20 MB or so at a time; this detail shouldn't matter) of the dirtied pages to the backing block device. Since the dirty pages are often rather interspersed in the region file, such a flush usually requires a couple of thousands of write ops, so it might sometimes take up to 10-20 seconds for the requests to complete.
    If the program, then, again tries to dirty any of the pages while they are waiting to be flushed, which is often the case, the VM will block it until the page in question is flushed. This means that the thread in question might very well be blocked for up to 20 seconds, causing quite annoying wait times.
    How to deal with this problem? I've considered trying to put the region files on tmpfs or so, but that seems like such an excessive measure for a problem which, from what I can tell, should be commonplace.
    On a very related note, I've noticed a large discrepancy in the I/O performance between the systems I've tried this on. Two of the systems in question manage to carry out some 200-500 write ops per second on my test load, while the third manages closer to 2000-3000 write ops per second, which makes quite a difference. What makes it very weird is that the faster system uses the exact same hard drive as one of the slower systems. I know this isn't exactly a BDB-specific question, but I thought someone around here might have experience in the matter. All three systems use Linux and S-ATA hard disks (not SSDs), but they use different S-ATA host adapters, different kernel versions and are configured in quite different ways.
    Thanks for reading my wall of text! I'm sorry for dragging on so long, but I didn't know how to describe the situation more briefly.
    Edited by: Dolda2000 on Mar 23, 2013 8:08 AM

    As a follow-up on this, it appears that the blocking behavior was introduced in Linux 3.0 to stabilize pages under writeback:
    http://lwn.net/Articles/486311/
    It seems that the commits that introduced the behavior can be safely patched away, and also that it is due to change in 3.9, but for now, this is not the route I took to solve it.
    Rather, I wrote a patch to Berkeley DB to allow me to store the region files in another directory than the environment root directory, and used it to store them in /dev/shm -- that is, on tmpfs, which avoids writeback of the region files altogether.
    If you want the patch, it is here for db4.8 (which what Debian Stable uses), and here for 5.1, which is what Debian Testing uses.
    (For some reason, the hyperlink format suggested by the forum doesn't seem to be working?)

  • YouTube Excessively Slow

    Over the past few weeks Youtube Videos have been excessively slow. I have a good cable connection to my Time Capsule. Every time I visit youtube the videos take an exceedingly long amount of time to download. Even for videos less than a minute long.
    I can game on these connections although occasionally there are bits of connection/packet loss. So It's not a connection speed issue. Other websites are working great.
    Why would this happen?

    I've had some interesting problems with this Click-to-Flash thing. While the concept is wonderful, I noticed a very significant degradation in performance when I actually did click on Flash that I wanted to see. In doing some packet capture on my machine, I can see that in many cases I'm somehow getting two copies of the same feed simultaneously. Even when I'm not, the sluggishness can be and almost always is extreme. A simple 30-second video might now require five or more minutes to buffer up sufficiently over a 1.5 Mbps DSL line. I just upgraded to 0.4 RC1 and we'll see if things improve...

  • Report Performance Issue - Activity

    Hi gurus,
    I'm developing an Activity report using Transactional database (Online real time object).
    the purpose of the report is to list down all contacts related activities and activities NOT related to Contact by activity owner (user id).
    In order to fullfill that requirment I've created 2 report
    1) All Activities related to Contact -- Report A
    pull in Acitivity ID , Activity Type, Status, Contact ID
    2) All Activities not related to Contact UNION All Activities related to Contact (Base report) -- Report B
    to get the list of activities not related to contact i'm using Advanced filter based on result of another request which is I think is the part that slow down the query.
    <Activity ID not equal to any Activity ID in Report B>
    Anyone encountered performance issue due to the advanced filter in analytic before?
    any input is really appriciated
    Thanks in advanced,
    Fina

    Fina,
    Union is always the last option. If you can get all record in one report, do not use union.
    since all records, which you are targeting, are in the activity subject area, it is not nessecery to combine reports. add a column with the following logic
    if contact id is null (or = 'Unspecified') then owner name else contact name
    Hopefully, this is helping.

  • Interested by performance issue ?  Read this !  If you can explain, you're a master Jedi !

    This is the question we will try to answer...
    What si the bottle neck (hardware) of Adobe Premiere Pro CS6
    I used PPBM5 as a benchmark testing template.
    All the data and log as been collected using performance counter
    First of all, describe my computer...
    Operating System
    Microsoft Windows 8 Pro 64-bit
    CPU
    Intel Xeon E5 2687W @ 3.10GHz
    Sandy Bridge-EP/EX 32nm Technology
    RAM
    Corsair Dominator Platinum 64.0 GB DDR3
    Motherboard
    EVGA Corporation Classified SR-X
    Graphics
    PNY Nvidia Quadro 6000
    EVGA Nvidia GTX 680   // Yes, I created bench stats for both card
    Hard Drives
    16.0GB Romex RAMDISK (RAID)
    556GB LSI MegaRAID 9260-8i SATA3 6GB/s 5 disks with Fastpath Chip Installed (RAID 0)
    I have other RAID installed, but not relevant for the present post...
    PSU
    Cosair 1000 Watts
    After many days of tests, I wanna share my results with community and comment them.
    CPU Introduction
    I tested my cpu and pushed it at maximum speed to understand where is the limit, can I reach this limit and I've logged precisely all result in graph (See pictures 1).
    Intro : I tested my E5-XEON 2687W (8 Cores Hyperthread - 16 threads) to know if programs can use the maximum of it.  I used Prime 95 to get the result.  // I know this seem to be ordinary, but you will understand soon...
    The result : Yes, I can get 100% of my CPU with 1 program using 20 threads in parallel.  The CPU gives everything it can !
    Comment : I put 3 IO (cpu, disk, ram) on the graph of my computer during the test...
    (picture 1)
    Disk Introduction
    I tested my disk and pushed it at maximum speed to understand where is the limit and I've logged precisely all result in graph (See pictures 2).
    Intro : I tested my RAID 0 556GB (LSI MegaRAID 9260-8i SATA3 6GB/s 5 disks with Fastpath Chip Installed) to know if I can reach the maximum % disk usage (0% idle Time)
    The result : As you can see in picture 2, yes, I can get the max of my drive at ~ 1.2 Gb/sec read/write steady !
    Comment : I put 3 IO (cpu, disk, ram) on the graph of my computer during the test to see the impact of transfering many Go of data during ~10 sec...
    (picture 2)
    Now, I know my limits !  It's time to enter deeper in the subject !
    PPBM5 (H.264) Result
    I rendered the sequence (H.264) using Adobe Media Encoder.
    The result :
    My CPU is not used at 100%, the turn around 50%
    My Disk is totally idle !
    All the process usage are idle except process of (Adobe Media Encoder)
    The transfert rate seem to be a wave (up and down).  Probably caused by (Encrypt time....  write.... Encrypt time.... write...)  // It's ok, ~5Mb/sec during transfert rate !
    CPU Power management give 100% of clock to CPU during the encoding process (it's ok, the clock is stable during process).
    RAM, more than enough !  39 Go RAM free after the test !  // Excellent
    ~65 thread opened by Adobe Media Encoder (Good, thread is the sign that program try to using many cores !)
    GPU Load on card seem to be a wave also ! (up and down)  ~40% usage of GPU during the process of encoding.
    GPU Ram get 1.2Go of RAM (But with GTX 680, no problem and Quadro 6000 with 6 GB RAM, no problem !)
    Comment/Question : CPU is free (50%), disks are free (99%), GPU is free (60%), RAM is free (62%), my computer is not pushed at limit during the encoding process.  Why ????  Is there some time delay in the encoding process ?
    Other : Quadro 6000 & GTX 680 gives the same result !
    (picture 3)
    PPBM5 (Disk Test) Result (RAID LSI)
    I rendered the sequence (Disk Test) using Adobe Media Encoder on my RAID 0 LSI disk.
    The result :
    My CPU is not used at 100%
    My Disk wave and wave again, but far far from the limit !
    All the process usage are idle except process of (Adobe Media Encoder)
    The transfert rate wave and wave again (up and down).  Probably caused by (Buffering time....  write.... Buffering time.... write...)  // It's ok, ~375Mb/sec peak during transfert rate !  Easy !
    CPU Power management give 100% of clock to CPU during the encoding process (it's ok, the clock is stable during process).
    RAM, more than enough !  40.5 Go RAM free after the test !  // Excellent
    ~48 thread opened by Adobe Media Encoder (Good, thread is the sign that program try to using many cores !)
    GPU Load on card = 0 (This kind of encoding is GPU irrelevant)
    GPU Ram get 400Mb of RAM (No usage for encoding)
    Comment/Question : CPU is free (65%), disks are free (60%), GPU is free (100%), RAM is free (63%), my computer is not pushed at limit during the encoding process.  Why ????  Is there some time delay in the encoding process ?
    (picture 4)
    PPBM5 (Disk Test) Result (Direct in RAMDrive)
    I rendered the same sequence (Disk Test) using Adobe Media Encoder directly in my RamDrive
    Comment/Question : Look at the transfert rate under (picture 5).  It's exactly the same speed than with my RAID 0 LSI controller.  Impossible !  Look in the same picture the transfert rate I can reach with the ramdrive (> 3.0 Gb/sec steady) and I don't go under 30% of disk usage.  CPU is idle (70%), Disk is idle (100%), GPU is idle (100%) and RAM is free (63%).  // This kind of results let me REALLY confused.  It's smell bug and big problem with hardware and IO usage in CS6 !
    (picture 5)
    PPBM5 (MPEG-DVD) Result
    I rendered the sequence (MPEG-DVD) using Adobe Media Encoder.
    The result :
    My CPU is not used at 100%
    My Disk is totally idle !
    All the process usage are idle except process of (Adobe Media Encoder)
    The transfert rate wave and wave again (up and down).  Probably caused by (Encoding time....  write.... Encoding time.... write...)  // It's ok, ~2Mb/sec during transfert rate !  Real Joke !
    CPU Power management give 100% of clock to CPU during the encoding process (it's ok, the clock is stable during process).
    RAM, more than enough !  40 Go RAM free after the test !  // Excellent
    ~80 thread opened by Adobe Media Encoder (Lot of thread, but it's ok in multi-thread apps!)
    GPU Load on card = 100 (This use the maximum of my GPU)
    GPU Ram get 1Gb of RAM
    Comment/Question : CPU is free (70%), disks are free (98%), GPU is loaded (MAX), RAM is free (63%), my computer is pushed at limit during the encoding process for GPU only.  Now, for this kind of encoding, the speed limit is affected by the slower IO (Video Card GPU)
    Other : Quadro 6000 is slower than GTX 680 for this kind of encoding (~20 s slower than GTX).
    (picture 6)
    Encoding single clip FULL HD AVCHD to H.264 Result (Premiere Pro CS6)
    You can look the result in the picture.
    Comment/Question : CPU is free (55%), disks are free (99%), GPU is free (90%), RAM is free (65%), my computer is not pushed at limit during the encoding process.  Why ????   Adobe Premiere seem to have some bug with thread management.  My hardware is idle !  I understand AVCHD can be very difficult to decode, but where is the waste ?  My computer want, but the software not !
    (picture 7)
    Render composition using 3D Raytracer in After Effects CS6
    You can look the result in the picture.
    Comment : GPU seems to be the bottle neck when using After Effects.  CPU is free (99%), Disks are free (98%), Memory is free (60%) and it depend of the setting and type of project.
    Other : Quadro 6000 & GTX 680 gives the same result in time for rendering the composition.
    (picture 8)
    Conclusion
    There is nothing you can do (I thing) with CS6 to get better performance actually.  GTX 680 is the best (Consumer grade card) and the Quadro 6000 is the best (Profressional card).  Both of card give really similar result (I will probably return my GTX 680 since I not really get any better performance).  I not used Tesla card with my Quadro, but actually, both, Premiere Pro & After Effects doesn't use multi GPU.  I tried to used both card together (GTX & Quadro), but After Effects gives priority to the slower card (In this case, the GTX 680)
    Premiere Pro, I'm speechless !  Premiere Pro is not able to get max performance of my computer.  Not just 10% or 20%, but average 60%.  I'm a programmor, multi-threadling apps are difficult to manage and I can understand Adobe's programmor.  But actually, if anybody have comment about this post, tricks or any kind of solution, you can comment this post.  It's seem to be a bug...
    Thank you.

    Patrick,
    I can't explain everything, but let me give you some background as I understand it.
    The first issue is that CS6 has a far less efficient internal buffering or caching system than CS5/5.5. That is why the MPEG encoding in CS6 is roughly 2-3 times slower than the same test with CS5. There is some 'under-the-hood' processing going on that causes this significant performance loss.
    The second issue is that AME does not handle regular memory and inter-process memory very well. I have described this here: Latest News
    As to your test results, there are some other noteworthy things to mention. 3D Ray tracing in AE is not very good in using all CUDA cores. In fact it is lousy, it only uses very few cores and the threading is pretty bad and does not use the video card's capabilities effectively. Whether that is a driver issue with nVidia or an Adobe issue, I don't know, but whichever way you turn it, the end result is disappointing.
    The overhead AME carries in our tests is something we are looking into and the next test will only use direct export and no longer the AME queue, to avoid some of the problems you saw. That entails other problems for us, since we lose the capability to check encoding logs, but a solution is in the works.
    You see very low GPU usage during the H.264 test, since there are only very few accelerated parts in the timeline, in contrast to the MPEG2-DVD test, where there is rescaling going on and that is CUDA accelerated. The disk I/O test suffers from the problems mentioned above and is the reason that my own Disk I/O results are only 33 seconds with the current test, but when I extend the duration of that timeline to 3 hours, the direct export method gives me 22 seconds, although the amount of data to be written, 37,092 MB has increased threefold. An effective write speed of 1,686 MB/s.
    There are a number of performance issues with CS6 that Adobe is aware of, but whether they can be solved and in what time, I haven't the faintest idea.
    Just my $ 0.02

  • RE: Case 59063: performance issues w/ C TLIB and Forte3M

    Hi James,
    Could you give me a call, I am at my desk.
    I had meetings all day and couldn't respond to your calls earlier.
    -----Original Message-----
    From: James Min [mailto:jminbrio.forte.com]
    Sent: Thursday, March 30, 2000 2:50 PM
    To: Sharma, Sandeep; Pyatetskiy, Alexander
    Cc: sophiaforte.com; kenlforte.com; Tenerelli, Mike
    Subject: Re: Case 59063: performance issues w/ C TLIB and Forte 3M
    Hello,
    I just want to reiterate that we are very committed to working on
    this issue, and that our goal is to find out the root of the problem. But
    first I'd like to narrow down the avenues by process of elimination.
    Open Cursor is something that is commonly used in today's RDBMS. I
    know that you must test your query in ISQL using some kind of execute
    immediate, but Sybase should be able to handle an open cursor. I was
    wondering if your Sybase expert commented on the fact that the server is
    not responding to commonly used command like 'open cursor'. According to
    our developer, we are merely following the API from Sybase, and open cursor
    is not something that particularly slows down a query for several minutes
    (except maybe the very first time). The logs show that Forte is waiting for
    a status from the DB server. Actually, using prepared statements and open
    cursor ends up being more efficient in the long run.
    Some questions:
    1) Have you tried to do a prepared statement with open cursor in your ISQL
    session? If so, did it have the same slowness?
    2) How big is the table you are querying? How many rows are there? How many
    are returned?
    3) When there is a hang in Forte, is there disk-spinning or CPU usage in
    the database server side? On the Forte side? Absolutely no activity at all?
    We actually have a Sybase set-up here, and if you wish, we could test out
    your database and Forte PEX here. Since your queries seems to be running
    off of only one table, this might be the best option, as we could look at
    everything here, in house. To do this:
    a) BCP out the data into a flat file. (character format to make it portable)
    b) we need a script to create the table and indexes.
    c) the Forte PEX file of the app to test this out.
    d) the SQL staement that you issue in ISQL for comparison.
    If the situation warrants, we can give a concrete example of
    possible errors/bugs to a developer. Dial-in is still an option, but to be
    able to look at the TOOL code, database setup, etc. without the limitations
    of dial-up may be faster and more efficient. Please let me know if you can
    provide this, as well as the answers to the above questions, or if you have
    any questions.
    Regards,
    At 08:05 AM 3/30/00 -0500, Sharma, Sandeep wrote:
    James, Ken:
    FYI, see attached response from our Sybase expert, Dani Sasmita. She has
    already tried what you suggested and results are enclosed.
    ++
    Sandeep
    -----Original Message-----
    From: SASMITA, DANIAR
    Sent: Wednesday, March 29, 2000 6:43 PM
    To: Pyatetskiy, Alexander
    Cc: Sharma, Sandeep; Tenerelli, Mike
    Subject: Re: FW: Case 59063: Select using LIKE has performance
    issues
    w/ CTLIB and Forte 3M
    We did that trick already.
    When it is hanging, I can see what is doing.
    It is doing OPEN CURSOR. But not clear the exact statement of the cursor
    it is trying to open.
    When we run the query directly to Sybase, not using Forte, it is clearly
    not opening any cursor.
    And running it directly to Sybase many times, the response is always
    consistently fast.
    It is just when the query runs from Forte to Sybase, it opens a cursor.
    But again, in the Forte code, Alex is not using any cursor.
    In trying to capture the query,we even tried to audit any statementcoming
    to Sybase. Same thing, just open cursor. No cursor declaration anywhere.==============================================
    James Min
    Technical Support Engineer - Forte Tools
    Sun Microsystems, Inc.
    1800 Harrison St., 17th Fl.
    Oakland, CA 94612
    james.minsun.com
    510.869.2056
    ==============================================
    Support Hotline: 510-451-5400
    CUSTOMERS open a NEW CASE with Technical Support:
    http://www.forte.com/support/case_entry.html
    CUSTOMERS view your cases and enter follow-up transactions:
    http://www.forte.com/support/view_calls.html

    Earthlink wrote:
    Contrary to my understanding, the <font face="courier">with_pipeline</font> procedure runs 6 time slower than the legacy <font face="courier">no_pipeline</font> procedure. Am I missing something? Well, we're missing a lot here.
    Like:
    - a database version
    - how did you test
    - what data do you have, how is it distributed, indexed
    and so on.
    If you want to find out what's going on then use a TRACE with wait events.
    All nessecary steps are explained in these threads:
    HOW TO: Post a SQL statement tuning request - template posting
    http://oracle-randolf.blogspot.com/2009/02/basic-sql-statement-performance.html
    Another nice one is RUNSTATS:
    http://asktom.oracle.com/pls/asktom/ASKTOM.download_file?p_file=6551378329289980701

  • Returning multiple values from a called tabular form(performance issue)

    I hope someone can help with this.
    I have a form that calls another form to display a multiple column tabular list of values(needs to allow for user sorting so could not use a LOV).
    The user selects one or more records from the list by using check boxes. In order to detect the records selected I loop through the block looking for boxes checked off and return those records to the calling form via a PL/SQL table.
    The form displaying the tabular list loads quickly(about 5000 records in the base table). However when I select one or more values from the table and return back to the calling form, it takes a while(about 3-4 minutes) to return to the called form with the selected values.
    I guess it is going through the block(all 5000 records) looking for boxes checked off and that is what is causing the noticeable pause.
    Is this normal given the data volumes I have or are there any other perhaps better techniques or tricks I could use to improve performance. I am using Forms6i.
    Sorry for being so long-winded and thanks in advance for any help.

    Try writing to your PL/SQL table when the user selects (or remove when deselect) by usuing a when-checkbox-changed trigger. This will eliminate the need for you top loop through a block with 5000 records and should improve your performance.
    I am not aware of any performance issues with PL/SQL tables in forms, but if you still have slow performance try using a shared record-group instead. I have used these in the past for exactly the same thing and had no performance problems.
    Hope this helps,
    Candace Stover
    Forms Product Management

  • Performance issues with class loader on Windows server

    We are observing some performance issues in our application. We are Using weblogic 11g with Java6 on a windows 2003 server
    The thread dumps indicate many threads are waiting in queue for the native file methods:
    "[ACTIVE] ExecuteThread: '106' for queue: 'weblogic.kernel.Default (self-tuning)'" RUNNABLE
         java.io.WinNTFileSystem.getBooleanAttributes(Native Method)
         java.io.File.exists(Unknown Source)
         weblogic.utils.classloaders.ClasspathClassFinder.getFileSource(ClasspathClassFinder.java:398)
         weblogic.utils.classloaders.ClasspathClassFinder.getSourcesInternal(ClasspathClassFinder.java:347)
         weblogic.utils.classloaders.ClasspathClassFinder.getSource(ClasspathClassFinder.java:316)
         weblogic.application.io.ManifestFinder.getSource(ManifestFinder.java:75)
         weblogic.utils.classloaders.MultiClassFinder.getSource(MultiClassFinder.java:67)
         weblogic.application.utils.CompositeWebAppFinder.getSource(CompositeWebAppFinder.java:71)
         weblogic.utils.classloaders.MultiClassFinder.getSource(MultiClassFinder.java:67)
         weblogic.utils.classloaders.MultiClassFinder.getSource(MultiClassFinder.java:67)
         weblogic.utils.classloaders.CodeGenClassFinder.getSource(CodeGenClassFinder.java:33)
         weblogic.utils.classloaders.GenericClassLoader.findResource(GenericClassLoader.java:210)
         weblogic.utils.classloaders.GenericClassLoader.getResourceInternal(GenericClassLoader.java:160)
         weblogic.utils.classloaders.GenericClassLoader.getResource(GenericClassLoader.java:182)
         java.lang.ClassLoader.getResourceAsStream(Unknown Source)
         javax.xml.parsers.SecuritySupport$4.run(Unknown Source)
         java.security.AccessController.doPrivileged(Native Method)
         javax.xml.parsers.SecuritySupport.getResourceAsStream(Unknown Source)
         javax.xml.parsers.FactoryFinder.findJarServiceProvider(Unknown Source)
         javax.xml.parsers.FactoryFinder.find(Unknown Source)
         javax.xml.parsers.DocumentBuilderFactory.newInstance(Unknown Source)
         org.ajax4jsf.context.ResponseWriterContentHandler.<init>(ResponseWriterContentHandler.java:48)
         org.ajax4jsf.context.ViewResources$HeadResponseWriter.<init>(ViewResources.java:259)
         org.ajax4jsf.context.ViewResources.processHeadResources(ViewResources.java:445)
         org.ajax4jsf.application.AjaxViewHandler.renderView(AjaxViewHandler.java:193)
         org.apache.myfaces.lifecycle.RenderResponseExecutor.execute(RenderResponseExecutor.java:41)
         org.apache.myfaces.lifecycle.LifecycleImpl.render(LifecycleImpl.java:140)
    On googling this seems to be an issue with java file handling on windows servers and I couldn't find a solution yet. Any recommendation or pointer is appreciated

    Hi shubhu,
    I just analyzed your partial Thread Dump data, the problem is that the ajax4jsf framework ResponseWriterContentHandler triggers internally a new instance of the DocumentBuilderFactory; every time; triggering heavy IO contention because of Class loader / JAR file search operations.
    Too many of these IO operations under heavy load will create excessive contention and severe performance degradation; regardless of the OS you are running your JVM on.
    Please review the link below and see if this is related to your problem.. This is a known issue in JBOSS JIRA when using RichFaces / ajaxJSF.
    https://issues.jboss.org/browse/JBPAPP-6166
    Regards,
    P-H
    http://javaeesupportpatterns.blogspot.com/

Maybe you are looking for

  • OBIEE 10g Dashboard Design- Embedding BI publisher report

    Hello all, i am trying for a specific layout functionality in the dashboard page. i want to display all the (BI publisher) report links in one section and upon clicking on each link the report should open in another section but in the same dashboard

  • IPhone 6 display got zoomed while using the camera (happened after update to iOS 8.1.3)

    Hi, after updating to the newest iOS version (iOS 8.3) I have experienced a lot of graphics issues. I experiensed it in the keyboard with some of the "button pop-ups" lingers. Also, when I rotate the phone it kinda shows a negative version of the scr

  • How do I set-up a group to text message?

    1st time user, still learning and getting used to this new iphone 4S. Am I able to set up and store groups to text message? I have @ 50 people to text message/email frequently and I used to do this with my old phone but I can't figure out how to do i

  • SHELF LIFE CONFIGURATION ISSUE

    Dear Guru I have an issue regarding shelf life material. I want that SLED should be calculate only for those material for which i have maintain shelf life data in material master. but  when i am trying to confirm process order for finished goods for 

  • Error when enable Reader Extended additional features

    Hi, I have created a form and look ok but when I enable Reader Extended PDF -> Additional Features form giving following error. How can I fix that? Regards Form