Slow Performance - Java Related?

This is an old box that I bought used recently, but the system install is recent. The system performs very slowly - about 50% of the speed of comparable Mac's in the XBench database. If I use Xupport to manully run "All" of the system maintenance crons I get some improvement, but it quickly goes back to being slow.
Looking at my logs I have a "boat load" of Java errors under CrashReporter; there will be a string of "JavaNativeCrash_pidXXX.log" entries - many of them, as follows:
An unexpected exception has been detected in native code outside the VM.
Unexpected Signal : Bus Error occurred at PC=0x908611EC
Function=[Unknown.]
Library=/usr/lib/libobjc.A.dylib
NOTE: We are unable to locate the function name symbol for the error
just occurred. Please refer to release documentation for possible
reason and solutions.
Many of the line entries that follow, but not all of them, refer to SargentD2OL, which is a Java app, which I installed, but it did not work properly so I removed it. Yet I continue to get Java errors that refer to this now non-existant app.
I have read that Java apps use a lot of resources, and that D2OL in particular uses a lot of resources. Can my slow performance problem be Java related? If so, any idea of how I can fix this problem?
G4 AGP Graphics   Mac OS X (10.3.9)   500 MHz, 512M RAM

Sorry to take so long to respond, but other issues in life have demanded my attention.
None of the solutions given have had any affect. My Java folder has both a 1.3.1 and a 1.4.2 app - the Java Update 2 will not reinstall because it sees an up-to-date app in the folder. But reading the update file it says the older Java will be removed - but it is still there. Problem?
On XBench the system scores a 9 to 10, while similar boxes on the XBench database score around 18 to 20. My cpu, memory, and video scores are very low. The HD through-put scores are the only ones that are normal. TechTool Pro 4 finds no problems. I have removed the memory sticks one at a time and retested after each cycle - no difference.
I have two drives, each with a 10.3.9 install. One works fine, scores around a 17 on XBench, the other scores a 9 to 10. So it appears to be a software problem. The slower install is a drive from a iMac G3 that has been moved to the G4 - are there issues with this?
My favored drive is the prior G3 one (newer and faster than the other drive that system tests faster in XBench) - it has my profile and all my info on it. It worked fine in the G3 - no problems.
Thanks for the help,
G4 AGP Graphics Mac OS X (10.3.9) 500 MHz, 512M RAM, ATI 8500

Similar Messages

  • Extremely slow performance with ojdbc6.jar on IBM JVMs (Java 6)

    We are consistently seeing slow performance (easily demonstrable by the simplest of test cases) while using ojdbc6.jar on IBM JDKs. Pefrormance is normal when we simply opt for ojdbc14.jar under the same JDK, same java program.
    Works well
    =======
    JRockit, ojdbc6.jar
    JRockit, ojdbc14.jar
    IBM JDK, ojdbc14.jar
    A lot slower
    =======
    IBM JDK, ojdbc6.jar
    All we had to do was to write a simple JDBC access program and measure time taken around
    Connection conn = DriverManager.getConnection(jdbcURL, dbProps);
    The first 3 combinations mentioned above return comparable numbers. The 4th combination takes 3-4 times as much.
    We initially saw this during performance test of our software on WebSphere (also supported on Weblogic). Later was able to eliminate WebSphere and Weblogic as the problem areas and narrow it down to the JVM. Reproduce-able on development machines.
    We have been using Oracle/WLS since BEA used to market 'Tengah' servers. This is the worst we have seen as far as the qaulity of support staff goes. We first level support seems clueless. We filed this case and the support comes back tells us that IBM JDKs are not supported for WLS! Oh, we have been running WLS/AIX/IBM JVM combination only for 5+ years.

    A quick solution to test this would be to do the following:
    mv /dev/random /dev/random.bk
    ln /dev/urandom /dev/random
    Note: The quick solution will disappear after a reboot. You could add it to rc.local to automatically make this change, but I wouldn't recommend it.
    The following is from [http://www.usn-it.de/index.php/2009/02/20/oracle-11g-jdbc-driver-hangs-blocked-by-devrandom-entropy-pool-empty/]:
    "Oracle 11g JDBC driver hangs blocked by /dev/random – entropy pool empty
    On a headless (=without console) network server, the 11g JDBC driver used for (java) application connect may cause trouble. In my case, it refused to connect to the DB without any error, trace or log entry. It simply hung. After several hours, it connected one time, and freezed again. Remote debugging done by the development clarified that it locks after calling SeedGenerator() and SecureRandom().
    Reason: The JDBC 11g needs about 40 bytes of secure random numbers, gathered from /dev/random, to encrypt its connect string.
    But public-available “man 4 random” says:
    When read, the /dev/random device will only return random bytes within the estimated number of bits of noise in the entropy pool. /dev/random should be suitable for uses that need very high quality randomness such as one-time pad or key generation. When the entropy pool is empty, reads from /dev/random will block until additional environmental noise is gathered.
    So far so good, now the question arises: Why does this mystic “entropy pool” runs out of gas?
    The answer is as simple as unsatisfying: because too less entropy “noise” was generated by the system. You can check the “filling level” (maybe zero?) of your pool and the overall size of the pool (usually 4096) by issuing
    cat /proc/sys/kernel/random/entropy_avail
    cat /proc/sys/kernel/random/poolsize
    Hint: /dev/random will deliver one new random number as soon as the pool has reached more than 64 entropy units.
    So why does my box not generate more entropy noise?
    Because only few drivers will fill the entropy pool, first of all keyboard and mouse. Sounds very useful on a server in a datacenter, isn’t it? Some block device and network drivers seem to do so as well, and I have read from guys on the net changing their network card and driver to enjoy this “feature”! But let’s stop ranting, /dev/random is simply made for high security randomness, and if it can’t make sure that randomness is as good as possible in this deterministic world, it stops. Intelligent people have created /dev/urandom for that, like “man 4 random” clearly states:
    A read from the /dev/urandom device will not block waiting for more entropy. As a result, if there is not sufficient entropy in the entropy pool, the returned values are theoretically vulnerable to a cryptographic attack on the algorithms used by the driver. Knowledge of how to do this is not available in the current non-classified literature, but it is theoretically possible that such an attack may exist. If this is a concern in your application, use /dev/random instead.
    Now let’s get back on our JDBC problem. Oracle JDBC 11g seems to use /dev/random by default, which causes usually no trouble on clients running with console access by a user, because his/her unpredictable :) actions will keep the entropy pool well-fed. But to make it usable on a headless server with a latently empty entropy pool, you should do several things, in descending security order (without warranty):
    1. Involve an audio entroy daemon like AED to gather noise from your datacenter with an open microphone, maybe combine it with a webcam noise collector like VED. Other sources are talking about “Cryptographic Randomness from Air Turbulence in Disk devices“. :)
    2. Use the Entropy Gathering Daemon to collect weaker entropy from randomness of userspace programs.
    3. Talk your JDBC into using /dev/urandom instead:
    -Djava.security.egd=file:///dev/urandom"

  • Slow performance of PreparedStatement

    I am having difficulty with extremely slow performance or a relatively simply Microsoft SQL Server 2000 call from my Java applet. Most of the application runs fine, but there are certain parts that are repeatedly giving me a long delay before completing their execution. The code is as follows:
    // Create the try block for the execution of the SQL code
            try{
                // Create the command to be executed
                String command = new String( "select Description, Enabled " );
                command += "from tLineInfo where( Line = ? )";
                // Create the SQL text to be executed
                PreparedStatement get = _connection.prepareStatement( command );
                get.setInt( 1, lineNo );
                // Execute the SQL command
                ResultSet lineInfo = get.executeQuery();
                // Display the information accordingly
                if( lineInfo.next() ){
                    // Populate the user data fields
                    description.setText( lineInfo.getString( "Description" ) );
                          <! more display code here >
    }   // End of if statment
                else{
                    // Clear the user data fields
                    description.setText( " " );
                           <! more display code here >
                }   // End of else statements
                // Close the result set in preparation for the next query
                lineInfo.close();
            }   // End of try block
            catch( Exception e ){
                // Display a dialog box informing the user of the problem
                Object[] options = { "     OK     " };
                JOptionPane.showOptionDialog( null, e.getMessage(),
                        "Error",
                        JOptionPane.OK_OPTION, JOptionPane.ERROR_MESSAGE,
                        null, options, options[ 0 ] ); 
            }   // End of Exception catch
            // Get the related area information
            populateAreaCombo( lineNo );       
    private void populateAreaCombo( int lineNo ){
            // Format the areaComboBox
             try{
                // Create a command to get the devices from the database
                String command = new String( "select Area, [Name], [Description] " );
                command += "from tAreas where( Line = ? )";
                // Create the SQL statement to grab the information, and
                // populate the search parameter
                PreparedStatement getAreas =
                        _connection.prepareStatement( command );
                getAreas.setInt( 1, lineNo );
                // Execute the command
                ResultSet areas = getAreas.executeQuery();
                // Loop through the result set, and add collect the areas
                Vector< String > controlAreas = new Vector();
                while( areas.next() ){
                    // Add the area to the comboBox
                    controlAreas.add( Integer.toString( areas.getInt( "Area" ) ) +
                            " - " + areas.getString( "Name" ) + ": " +
                            areas.getString( "Description" ) );
                }   // End of while loop
                      <! more display code here >
                    The application always seem to pause at the second PrepraredStatement call:
    PreparedStatement getAreas =
                        _connection.prepareStatement( command );This seems to be a very simple operation, and it is not even the execution of the query where the long delay is realized. Rather, it is in the actual creation of the object prior to the execution.
    The delay is very repeatable at this exact statement each time.
    Additionally, of interest, is that the delay is only realized on computers remotely connected to the database. If I run this code on the localhost, then there is no delay. As soon as I distribute it, then the delay is incurred. That being said, there is not a network related issue that I can identify here. I have even isolated the server to be on the network with just one other PC, and the delay still persisted.
    Does anyone have any ideas?
    Thanks

    I can determine where the delay occurs by adding
    dialog boxes at a bunch of different steps, then
    monitoring them for when they appear; a little
    archaic, of course, but an easy way to find this
    out.You should
    1. Get the start time
    2. Get the current time at each step.
    3. Print the results at the end.
    4. Repeat a number of times to average.
    >
    When you say that I should not mix database code with
    display code... What exactly do you mean? To be
    more precise in my description, I was simply setting
    a bunch of different text fields and/or check boxes,
    etc., based on the result set returned. It was not
    as if I was creating a portion of the GUI there or
    something. I am assuming that is an allowable
    practice...
    You should have a class that does nothing but the database work. That class should be used by other classes (like classes that do GUI.)
    I have made little effort to close my resources, and
    sometimes they are not closed at all. When you say
    "resources", what exactly do you mean by that? Are
    you referring to the result sets, for example? I do
    not know of other resources that need to be closed,
    except for the connection to the DB itself. This, I
    have as persistent throughout the duration of the
    user's session.
    You must close result sets, statements and connections. They must be closed in that order.
    I am not running this code on the Internet, but it
    has been designed to be run on a small corporate
    network ( < 10 users). This is why I opted for the
    applet to run the entire application through instead
    of doing more HTML work.That is ok.

  • Linux AMD64, JDK 1.5_03: slow performance with large heap

    Tomcat app server running on jdk 1.4.2 on 32 bit Linux configured with mx1750m, ms1750m, runs fast. Returns 2MB of data through HttpServlet in under 30 secs.
    Moving the same app server to 64 bit on jdk 1.5.03 configured with mx13000m, ms10000m, the same request for data takes 5-20 minutes. Not sure why the timing is not consistent. If the app server is configured with mx1750m, ms1750m, performance is about 60 secs or less.
    I checked java settings through jstat. -d64 is the default. Why would increasing the heap cause such slow performance? Physical memory on the box = 32MB.
    It looks like it's definitely java related since a perl app making a http request to the server takes under a minute to run. We moved to 64bit to get around 1.7GB limitation of 32bit Linux but now performance is unacceptable.

    I Aggree, a AMD 64 with only 32 MB of memory would be a very strange beast indeed, heck, my graphics card has 4 times that, and it's not the most up-to-date.
    Keep in mind that switching to 64 does not only mean bigger memory space but also bigger pointers (on the sub-java level) and probably more padding in your memory, which leads to bigger memory consumption which in turn leads to more bus traffic which slows stuff down. This might be a cause for your slowdown, but that should not usually result in a slowdown as sever as the one you noticed.
    Maybe it's also a simple question of a not-yet-completely-optimized JDK for amd64.

  • Slow Performance Forms 10g !!!!

    Hi,
    I have migrated several forms from 6i to 10g and i use a 11g Database.
    Most of the users accessing from different countries complain of slow performance but the basic idea of migration is to get a better performance. My users mostly work on laptops.
    I am unable to find where the problem is. Is this might be due to network traffic since they are accessing via internet or any server related issues. Please help me to find and fix the problem and is there any tips for gaining better performance in using forms 10g.
    Regards,
    Suresh

    So, were they running on 6i via laptop and accessing via the internet before you migrated ? (ie: was it 6i client server or 6i web forms)
    Did you migrate the database too ?
    Slow performance as in : it used to be seconds and now it's taking minutes ?
    Steve

  • Slow performance on Intel Mac a damaged disk?

    Hi,
    I'm experiencing very slow performance on my Intel Mac 216 Ghz core duo, with 2GB 667 MHz DDR2 SDRAM. Running OS 10.6.8. Have been experiencing reduced performance, can't use Entourage email or Chrome, Firefox browsers at the same time. Ran Utilities to repair permissions, only a Java set of issues. Disk Utilities found no problems with the disk.  Have experienced this issue increasingly for 5-6 months. Using Activity Window i see that Entourage, Word, Safari, Chrome, even System Preferences pane sometimes "Is not responding". Other system services take longer than before.
    Apple support had me get larger RAM chip. Now have 2GB Ram. Upgraded to OS 10.6.8 with no improvement. Ran Software Update several times.
    I have no anti virus software, have disabled Spotlight and any other memory demanding software. I'm on a in-home wi-fi network on an Airport Extreme.
    In addition, at startup it takes 15+ seconds for the keyboard (wired) typing to show in the log on field. Running a single account with Admin access. Screen refreshes slow at times. Currently running is Safe Boot mode but that's not helping much.
    Read on this board discussion 23238910 to use Terminal Window to do a syslog comand. The results in the Text Edit reads
    Thu Jun  5 14:29:54 strait-admins-computer kernel[0] <Debug>: disk1: I/O error.
    Thu Jun  5 14:32:25 strait-admins-computer kernel[0] <Debug>: disk1: I/O error.
    Thu Jun  5 16:44:10 strait-admins-computer kernel[0] <Debug>: disk1: I/O error.
    Fri Jun  6 09:07:00 strait-admins-computer kernel[0] <Debug>: disk1: I/O error.
    I need to resolve the issue as it's killing my time!
    Thanks for any suggestions.

    It does not sound good for your disk drive.
    I suggest you backup everything to an external disk drive ASAP.
    Review the steps to changing the disk drive at http://www.ifixit.com/Guide/iMac+Intel+20-Inch+EMC+2266+Hard+Drive+Replacement/9 19

  • Slow performance of JDBC - ODBC MS ACCESS

    I experience a very slow performance for jdbc-odbc using ms access as the database. This program works fine with other computer (in term of performance). However, the harddrive is cranking big time with this computer (this is the fastest one among the computers I tested, and also has many gigabytes left to be used). The database is very small. Other computer use exactly the same java version and msaccess driver version. If anyone found the same problem, or have any suggestion please help. Thank you.

    I am having the same problem with one machine as well. Running MS Access 2000 (unfortunately), and all machines run well with one exception. DB reads take about 10 seconds each. If a solution has been found, please report.
    --Dave                                                                                                                                                                                                                                                                                                                                                                                                                                                                           

  • Slow Performance or XDP File size very large

    There have been a few reports of people having slow performance in their forms (tyically for Dynamic forms) or file sizes of XDP files being very large.
    These are the symptoms of a problem with cut and paste in Designer where a Process Instruction (PI) used to control how Designer displays a specific palette is repeated many many times. If you look in your XDP source and see this line repeated more than once then you have the issue:
    The problem has been resolved by applying a style sheet to the XDP and removing the instruction (until now). A patch has been released that will fix the cut and paste issue as well as repair your templates when you open them in a designer with the patch applied.
    Here is a blog entry that describes the patch as well as where to get it.
    http://blogs.adobe.com/livecycle/2009/03/post.html

    My XDP file grow up to 145mb before i decided to see what was actually happening.
    It appears that the LvieCycle Designer ES program sometimes writes alot of redundant data... the same line millions of times over & over again.
    I wrote this small java program which reduced the size up to 111KB !!!!!!!!!!!!!!!!!! (wow what a bug that must have been!!!)
    Here's the sourcecode:
    import java.io.BufferedReader;
    import java.io.BufferedWriter;
    import java.io.FileNotFoundException;
    import java.io.FileReader;
    import java.io.FileWriter;
    import java.io.IOException;
    public class MakeSmaller {
    private static final String DELETE_STRING = "                           <?templateDesigner StyleID aped3?>";
    public static void main(String... args) {
      BufferedReader br = null;
      BufferedWriter bw = null;
      try {
       br = new BufferedReader(new FileReader(args[0]));
       bw = new BufferedWriter(new BufferedWriter(new FileWriter(args[0] + ".small")));
       String line = null;
       boolean firstOccurence = true;
       while((line = br.readLine()) != null) {
        if (line.equals(DELETE_STRING)) {
         if (firstOccurence) {
          bw.write(line + "\n");
          firstOccurence = false;
        } else {
         bw.write(line + "\n");
         firstOccurence = true;
      } catch (FileNotFoundException e) {
       e.printStackTrace();
      } catch (IOException e) {
       e.printStackTrace();
      } finally {
       if (br != null) {
        try {
         br.close();
        } catch (IOException e) {
         e.printStackTrace();
       if (bw != null) {
        try {
         bw.close();
        } catch (IOException e) {
         e.printStackTrace();
    File that gets generated is the same as the xdp file (same location) but gets the extension .small. Just in case something goes wrong the original file is NOT modified as you can see in the source code. And yes Designer REALLY wrote that line like a gazillion times in the .xdp file (shame on the programmers!!)
    You can also see that i also write the first occurrence to the small file just in case its needed...

  • Bug report: Energy Manager causes slow performance...

    Very slow Windows Startup..
    Energy Manager's plans and the problem occured using any plan as I remember, and it wasn't an issue with the maximum processor state. I had heavy disk usage, especially access to the Page File, which is what the person in the second link I posted also described. CPU usage was low. The disk usage was seemingly not related to Energy Manager (Chrome was usually the worst), but removing Energy Manager completely resolved the problem. If you would find it helpful. I'm quite confident that even with Maximum Processor state at %100, CPU usage was low when it was slow, and that disk access was the culprit.
    previously have use 1.0.0.31 or 32 not sure.
    Windows 8.1
    Im using Lenovo Energy Management Driver
    1.0.0.28
    4.6.2014

    I actually have the same problem.  Sorry to work on a old thread, but...
    I've got my Lenovo B50-70 less than 1 month ago, and had switched it to SSD and had set everything up and is working perfectly.  However there are a bunch of lenovo programs that came pre-installed.  This computer is used mainly for work and I really don't like programs that suddenly pop-up and ask you to do stuff, as it is quite distracting, and with Lenovo Energy Manager being one of them so I removed it.  (I did do some research before I remove...)
    Everything was fine, until when I try to switch on my wifi (I usually have it off, as I use LAN most of the time)...  For the B50-70 there is a airplane button on F7, which allow the user to switch wifi and bluetooth on/off.  That button was unresponsive...  After some digging it is actually related to the power energy manager being removed.  So I've downloaded the latest version from Lenovo and it was working again...
    However though, whenever I start my computer it is taking so long to start...  And I tried uninstalling the power energy manager and it solved the problem...  And I am stuck with either having no wifi on/off ability or a really slow boot, which loses the whole purpose of the SSD.  After you enter your username and password it originally takes around 5~10 seconds to login, now it takes over 1 minute.
    I have found this other link, which says Lenovo Power Manager having the same problem.
    https://forums.lenovo.com/t5/ThinkVantage-Technologies/Bug-report-Power-Manager-causes-slow-performa...
    There is a solution, which you can change the registry to switch a function off.  I've tried it and it didn't work, which is not surprising as it's a different program..  But I am hoping that Lenovo can provide a similar solution for Lenovo Energy Manager...

  • Performance issues -- related to printing

    Hi All,
    I am haviing production system performance issues related to printing. endusers are telling the printing is slow for almost printers. We are having more that 40 to 50 printers in landscape.
    As per my primary investigation I didnt find any issues in TSP01 & TSP02 tables. But I can see the table TST01 and TST03 table having many number of entries (more that lakh). I dont have idead about this table. Is ther eany thing related to this table where the print causes slowness or any other factors also makes this printing issue .. Please advice ..
    thanks in advance

    Hai,
    Check the below link...
    http://help.sap.com/saphelp_nw70/helpdata/en/c1/1cca3bdcd73743e10000000a114084/content.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/fc/04ca3bb6f8c21de10000000a114084/frameset.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/86/1ccb3b560f194ce10000000a114084/content.htm
    TemSe cannot administer objects that require more than two gigabytes of storage space, regardless of whether the objects are stored in the database or in the file system. Spool requests of a size greater than two gigabytes must therefore be split into several smaller requests.
    It is enough if you perform the regular background jobs and Temse consistency checks for the tables.
    This will help in controlling the capacity problems.
    If you change the profile parameter rspo/store_location parameter value to 'G' this will make the performance better. The disadvantages are TemSe data must be backed up and restored separately from the database using operating system tools, In the event of problems, it can be difficult to restore consistency between the data held in files and the TemSeu2019s object management in the database. Also you have take care of the Hard disk requirements because in some cases, spool data may occupy several hundred megabytes of disk storage. If you use the G option, you must ensure that enough free disk space is available for spool data.
    Regards,
    Yoganand.V

  • Acrobat very Slow performance (50% CPU usage)

    My acrobat 9 PRO (9.3.2) has suddently started to use 50% of my CPU usage everytime a *.pdf document is opened.
    It is the task Acrobat.exe which is the problem. This results in VERY slow performance and I cannot use the review function without adobe is crashing!
    This has not always been a problem!
    Does anyone know why this is happening?
    I have tried to uninstall and reinstall acrobat, without any luck!
    Thanks
    Peder

    You probably have the same problem as I do. I use version 9.3.3 on a fairly powerful machine with Windows vista. In my case, the problem is related to the weblink.api (http://forums.adobe.com/thread/675941?tstart=0). I also uninstalled, reinstalled, repaired - nothing helps.

  • Powerbook painfully slow performance

    Recently my powerbook has been running unusably slow. Here's a timeline:
    - Noticed slow opening of applications, saving, switching from application to application, browsing folders through finder.
    - Computer kept freezing after 1-10 minutes of use.
    - Checked RAM in system profiler - 1GB RAM showing 512 MB.
    - Removed RAM and restarted (new RAM being sent to me at present).
    - Freezing problem stopped. Slow performance improved for 2-5 minutes after startup but after 10 minutes, back to same old slow usage.
    - Checked hard drive, repaired permissions, did macjanitor cleanup.
    - Hardware check from startup DVD showed no problems. SMART check OK'd hard drive.
    - Archive and re-install tiger and updated to 10.4.9.
    - Activity monitor shows %CPU usage as fairly normal. Kernel_task is using about 720MB of virtual memory under very light/no user usage.
    - Created a new user and performance normal in new user, leading me to believe that the problem is related to user library and possibly preferences.
    - Put preferences folder on desktop and started putting my preferences back into user library folder bit by bit, restarting and testing. The performance would be better each time but hard to test as I have to wait at least 10 mins to test each time - time consuming business.
    Is there any way to narrow down which preference file is causing the problem without painstakingly re-adding preference files one by one to the user library folder, waiting 10 mins and testing?
    Otherwise, could the diagnosis above be due to something else all together. Been through the discussions and can't get anything to work so far. The RAM was the cause of the freezing problem, but seems to have brought on the slowness problem.
    Any help would be much appreciated, thanks Graeme.

    Graeme:
    If you have not done so recently run Unix Maintenance (CRON cleaning). Download and install MacJanitor and run all tests.
    Dr. Smoke's FAQ Tuning Mac OS X Performance has some hints you may find helpful.
    Regualar maintenance also helps performance. Take a look at the following articles:
    Mac OS X 10.3/10.4: System maintenance
    Macintosh OS X Routine Maintenance
    Maintaining OS X
    Good luck.
    cornelius

  • Big Library:Slow Performance?

    I have about 10,000 photos (.jpg ~3MB each) in my Aperture Library. I want to add many more photos but I'm worried about slow performance. My questions:
    1. How big is your library or how big do you think the libraries can get?
    2. Have you noticed slower performance with larger libraries?
    3. Any opinion on breaking up into multiple smaller libraries vs. 1 larger library?

    I am running two libraries,
    one for all of my work related imagery, 15,000+ images 50/50 raw & jpegs and
    the other for all of my stuff I shoot of sporting clubs, the bit I give back to the community, 18,000+ predominantly jpeg's
    both run smoothly, one on the MacPro and the other on the G4 laptop.
    Issue starts to be the backing up, if you are thinking it will get BIG, try a library for each client. Could be a good selling point as well, "your imagery is isolated from other clients and has its own dedicated backup".
    Tony

  • OCFS slow performance in ImpDP

    Did anyone encounter slow performance using ImpDP on OCFS ? We have Oracle RAC 10.2g running in window connecting SAN storage and was formatted to OCFS. We compared the performance of running the same import in Local Drive (NTFS) and SAN storage (OCFS) and found that the time required to complete the import is 10 times than Local Drive (NTFS). Any tuning or troubleshooting that we could do ? Thanks in Advance.

    Not aware of ocfs installing automatically when installing CRS/RAC. But then have never used Windows as a production o/s for Oracle. never mind using it to run RAC.
    ocfs2 is the latest version as per Oracle's [Open Source project list|http://oss.oracle.com/projects/]. Am not aware of "ocfs 9" for 9i or "ocfs 10" for 10g on Windows.
    What do you use ocfs for? Do you have the OCR and voting disks on the file system as files? Do you use it for Oracle datafiles too?
    As for the SAN infrastructure being "+okay+" - there's more than that to it. You need some kind of driver on the o/s to interface with the HBA card.. where cables (dual 2Gb fibre?) run into a SAN switch and where the SAN switch is connected to the physical storage system.
    A faulty cable can cause performance issue (e.g. reducing the dual path to the SAN via 2 fibre channels to a single one). There could be errors on the switch. The driver could be problematic (older version). Etc.
    So there is a number of layers, both s/w and h/w, to check before going "+okay+". Trust that your infrastructure team is jacked up in this regard.
    One method to benchmark the import problem is to use NT's performance monitor and monitor the import process. Not exactly sure how you would do it dealing with threads (as a thread in the oracle.exe executable will be servicing the import process) - but this should give you a somewhat accurate figure of the I/Os per second and bytes and blocks per second that the process is reading from the ocfs file system (assuming that you are writing it via ASM to raw devices).
    At the same time you can monitor the process for other potential performance bottlenecks.
    The only other factor that comes to mind that can impact ocfs performance is the ocfs cluster's communication layer. Like CRS that uses Interconnect for communication between nodes, so does ocfs. On a Unix system this is configured in the +/etc/ocfs2/cluster.conf+ file - and lists the number of ocfs nodes in the cluster and the network interfaces/addresses used for communication.
    If this network layer is already quite busy, using it for ocfs2's cluster communication could be a problem performance wise.
    You can test this aspect by unmounting ocfs2 (and stopping the ocfs2 cluster communication/heartbeat s/w too) on all but one node and then running the import. As there is only a single node in the ocfs2 cluster, the communication with other nodes is pretty much made redundant.
    Last comment - ocfs2 is not slow as a cluster file system. It does a pretty decent job. And its performance is directly related to the h/w it runs on, and also how it is configured. That said, it is not a s/w component that really plays a role in Oracle CRS and RAC as ASM should be used (and is the recommended option from Oracle). ocfs2 at best can serve as a cooked file system for the RAC cluster (only used via directory objects in Oracle).

  • Windows Server 2012 very slow performance executing files

    Hello,
    I am running windows server 2012 as dc with 2 clients connected running win7 pro x64, this is a new setup.
    Whenever i try to open a large file from the client or an application from the client it takes forever to startup or to load. The performance monitor indicates a maximum workload of 15% at all times.
    I have tried all the following potential solutions i found, however none of them seem to do the trick.
    - enable write cache = http://social.technet.microsoft.com/Forums/en-US/a76f6e97-3266-4ba1-ba90-ad5318ec4937/slow-file-sharing-performance-on-windows-server?forum=winservergen
    - change group policy = http://support.microsoft.com/kb/829700/nl
    - change group policy as = http://social.technet.microsoft.com/Forums/en-US/bcb4b580-50ef-4f10-bc9b-16db1bf24804/slow-network-file-execution-on-windows-server-2012?forum=winserverfiles
    Does anyone have any ideas?
    With kind regards

    Hi,
    You could disable SMBv3 on server 2012 to check if the issue related to SMB protocol. 
    How to enable and disable SMBv1, SMBv2, and SMBv3 in Windows Vista, Windows Server 2008, Windows 7, Windows Server 2008 R2, Windows 8, and Windows Server 2012
    http://support.microsoft.com/kb/2696547/en-us
    Warning: We do not recommend that you disable SMBv2 or SMBv3. Disable SMBv2 or SMBv3 only as a temporary troubleshooting measure. Do not leave SMBv2 or SMBv3 disabled.
    Please also refer to the Forum FAQ to troubleshoot the slow performance issue:
    [Forum FAQ] Troubleshooting Network File Copy Slowness
    http://social.technet.microsoft.com/Forums/windowsserver/en-US/7bd9978c-69b4-42bf-90cd-fc7541ccb663/forum-faq-troubleshooting-network-file-copy-slowness?forum=winserverPN
    Regards,
    Mandy
    We
    are trying to better understand customer views on social support experience, so your participation in this
    interview project would be greatly appreciated if you have time.
    Thanks for helping make community forums a great place.

Maybe you are looking for

  • I have the same problem w/ 2 different vcams

    Okay peoples, I'm putting together an animation in Flash CS4, on a MacBook Pro running OS X 10.5.8, and I positively need to be able to use the vcam (of course!).  But I can't make it work. Essentially, when I try to test or publish the movie w/ the

  • Use XPI_Inspector for HTTP_AAE receiver

    Hi Experts, Current scenario: - I've created a synchronous Proxy to HTTP_AAE scenario which uses the Google Distance Matrix API Problem: - The response mapping gives the following error: "Values missing in queue context." Question: - I'm wondering if

  • Safari with Adblock crashing MCB Pro with Yosemite?

    I have an early 2013 Macbook Pro Retina / 2.6 GHz / 8 GB memory / Yosemite 10.10.1 Having upgraded to Yosemite in early November the MCB froze and eventually crashed a few times while on Safari but on restarting everything worked fine. 3 weeks later

  • SQ01 - ABAP List Report Header

    We have a query in SQ01 that provides a listing of employees and their IT0027 distribution accounts. Using the ABAP List as the output I have some control over the header information. I've done things like the report title, the username of the indivi

  • Itunes using excessive cpu

    I upgraded the iTunes to 11.0.4 Everytime it is running since then, even in the background i noticed that the process for it is has a vey high cpu usage (above 95% at all times). This is while im not even using it to play or sync or anything.. the pr