Cache Flushes Solaris10 StorageTek D280 and NFS and ZFS

I encounter complains from users, who are connected via nfs, to Sun Solaris10 server.
The server is connected via Fibre to a Storage Tek d280.
The performance on the server is okay.
However, on the , via nfs connected , clients, the performance is poor.
I found this document, and want to try to disable the cache flushes on the server.
http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide#Cache_Flushes
However, I rather want to have the Storage Tek D280 acting as a nice zfs storage device, rather than tweaking the operating system.
But I can not find any document on how to configure the behavior of the cache flushes on this device.
Is there someone who know how to setup this storage tek D280 box correctly to ingnore the Cache Flush commands generated by the NFS ?
Kind regards,

806760 wrote:
Thanks for the response.
I don't know how the D280 internally has been setup. It should use raid 5. That's about the only thing I know about it.
It is under control of an ICT department.
However, the effect, or if the D280 is poor configured, does it only affect the NFS clients connected to the Solaris Server ?
I have ruled out the network configuration. This is a 1Gb connection. And for diagnose I tryed with a different switch, and direct connection.
But it did not influence the poor performance of the client, using NFS.
As a test, I just extract a tar file with a big amount of empty files.
This goes over 25 times slower on the clients, than on the server.
I have installed about 8 of those systems, but none is performing so bad.
Since everything on all systems is about the same configuration, the only things which are out of my control, is the network, and the san.
I tryed to test the network, but I don't see any problems with that.
So in my mind, the only thing left, would be the san device.
Searching on this topic, I found some explanations, about the zfs with nfs, which works poorly, due to the nfs is committing a regular synchronous write (NFS commit) However I don't like to do this.
I also can not find any description on how to configure a D280.
It would be nice, if you could provide some settings which has to be set in a D280.
The configuration is two cluster nodes, and two clients.
The cluster node mainly task is to provide the nfs shares.
The clients and servers are in one 19" rack.
The San, I don't know where it is.
It has a 2Gb fibre coupling. ( On the server side there are 4Gb Emulex HBA's installed )
Kind regards,If a tar file extracts 25 times faster on the server then it does over the network, yet both times the data is being written to the SAN LUNs on the D280, the problem is the network.
That tar file extracts slower across the network for two reasons: bandwidth and latency.
There's only so much data you can stuff through a gigE network. Your single 1 gigE link can handle about 100 MB/sec read and 100 MB/sec write combined - total. For all users. That may be part of your performance problem, because the configuration LUN layout of that D280 would have to be really, REALLY bad for it to be unable handle that relatively small amount of IO. You CAN test the performance of the LUNs being presented to your server - just use your favorite benchmarking tool to do various reads from the "/dev/rdsk/...." device files that make up your filesystem(s). Just make doggone sure you ONLY do reads - if you write to those LUNs your filesystem(s) will be corrupted. Something like "dd if=/dev/rdsk/... of=/dev/null bs=1024k count=10000" will tell you how fast that one LUN can stream data - but it won't tell you how many IO ops/sec the LUN can support as you'd need to do random small reads to do that. Any halfway-decently configured D280 LUN should be able to stream data at a constant 200 MB/sec while you're reading from it.
And even if the bandwidth were much higher, you still have to deal with the additional latency of having to do all communications across your network. No matter how fat the pipe is, it still takes more time to send data across the network and have to wait for a reply. What do your ping times look like between client and server? And even with that added latency, there are some things you can do on your hosts. Increase your TCP buffer sizes, mount your filesystems on your Linux clients with the "rsize=32768,wsize=32768,intr,noatime" options, and maybe use NFSv3 instead of NFSv4 - make sure you change both the server and client settings to be sure. And work with your network admins to get jumbo frames enabled. Moving more data per packet is a good way to address latency because you wind up having to wait for a response much fewer times.

Similar Messages

  • SSPU issue: Cache directory does not exist, and couldn't create a new downl

    I have installed SSPU and created an offer over there. When I run my offer it gives the error:
    Cache directory does not exist, and couldn't create a new download cache directory.
    Can anyone tell me the root cause and workaround for this issue?

    Hi
    Try deleting the "offer" directory from SSPU home and then delete the offer from Sub-Client . Restart client and then add the provider . Start pushing the content from SSPU and see if the issue still persists .
    Thanks
    Srinath

  • Use cached exchange mode for new and existing outlook profiles

    Hi, I've noticed a difference in the behavior of this setting in Outlook 2010 vs 2013
    In 2010, when the gpo is set, everything is greyed out for the user, but in 2013, the user can change the setting. Has anyone else noticed this? Is this something that MS decided to change or a bug?
    I'll attach screenshots of the settings and the end result later.
    joeblow

    Hi,
    à
    In 2010, when the gpo is set, everything is greyed out for the user, but in 2013, the user can change the setting.
    Based on your description, I understand that “Use Cached Exchange Mode” will gray out in Outlook 2010 when
    set "Use Cached Exchange Mode for new and existing Outlook profiles".
    However, in Outlook 2013, you will still be able to check “Use Cached Exchange Mode”. If anything I misunderstand, please don’t hesitate to let me know.
    Just my guess, Office 2013 Administrative Templates may be a little difference with Office 2010 Administrative
    Templates. On current situation, I suggest that you would post the question in
    Office Forum. I believe we will get a better assistance there.
    Hope this helps.
    Best regards,
    Justin Gu

  • Putting a row into a pageable cache failed with error 9 and error message

    Hi Guys,
                 I am fetting the follwoing message when the job is scheduled.
    Putting a row into a pageable cache failed with error <9> and error message <Db::put: Bad file descriptor>
    I read the forum and Manoj has suggested to one of the forum member to change the size of the Pcache setting through the server manager. But I do not see the setting there. can any one please let me know where exactly I can see and increase the size of the Pcache.
    Thanks

    Hi
    I've had the same problem.  We know from other (more meaingful) error messages that we ran out of space for the Pageable Cache, but we've just increased it by 50GB and still get the same error.
    Anyone??
    Btw i don't think you there is a setting for PCache size - it will just use all the free disk you have on the drive where you have located the PCache directory.
    Richard

  • Trying for 3 days to get Bridge to work. Told me to unable ti read cache -- tired everything including deleting and reloading -- still getting same message. Pl;ease help I am on a deadline. Bob

    Bridge unable to read cache
    Trying for 3 days to get Bridge to work. Told me to unable ti read cache -- tired everything including deleting and reloading -- still getting same message. Restarted computer several times as well. Please help I am on a deadline. Bob

    THANKS FOR REPLY,
    No, I am on a Mac  OS X 20.9.4
    BOB

  • My safari screen is greyed out, I checked and found a response to this question.  I went to settings and cleared history, cookies, and cache, java script is on and I restarted, screen is still grey, cant touch it to do a thing.  Help!

    my safari screen is greyed out, I checked and found a response to this question.  I went to settings and cleared history, cookies, and cache, java script is on and I restarted, screen is still grey, cant touch it to do a thing.  Help!

    That is usually the fix but try this now.
    Reset the iPad by holding down on the sleep and home buttons at the same time for about 10-15 seconds until the Apple logo appears - ignore the red slider - let go of the buttons. See if that helps.
    You can also try this - quit Safari. From the home screen - Double tap the home button and your recent apps appear the bottom. Tap and hold down on the Safari icon until it wiggles. Tap the minus sign in the upper left corner to close the app. Tap the home button twice. Restart the iPad.
    Message was edited by: Demo

  • Cache Configuration Assistant goes on and on and on

    Hi all,
    I am installing Oracle 9ias 1022 on a windonw 2000 machine. The origin databas (8i 81701) is on a Linux box.
    The installation seems to have stuck at the Cache Confoguration Assistant Progress Dialog window. Its been at 16% for the past two hours! Is that normal? can I cancel and continue?
    Any pointers will be most welcome.
    Keita.

    I have specifically not installed on Linux database, but the iCache configuration hanging at 16% is reported a couple of times which is mainly due to the pre-configuration required for database.
    If you are not really keen on using database cache you can cancel it and go ahead.
    If you want to use database cache, from the cmd prompt use
    cmd> tnsping extproc_connection_data
    and see if it returns ok.
    If not that means you have not configured your database. Refer to preinstalltion steps in installation guide and then invoke the cache configuration assistant using the cmd
    cmd > wtacca -create -typical
    null

  • AAA user cache flush not working after 10.1.4.2 patch

    After we applied the OAM 10.1.4.2 patch to our test and QA environments, user cache flushes don't appear to work on the Access server. This is the case for both automatic flushes via the Identity server, and manual flushes from the Access server console. There aren't any errors displayed, but the user data isn't updated in the cache.
    The services have all been restarted, and the Identity server settings "doAccessServerFlush" and "enableAllowAccessCache" are still set to true.
    Has anyone else seen this issue? Any ideas?

    This is an old issue, but I finally got the last part of the solution today from Oracle. The solution is to update the "UserMgmtNodeEnabled" setting in the globalparams.xml files for both the Access Server and the Policy Manager. This value was added by the 10.1.4.2 update, and it is set by default to "False". It should be set to "true" to fix the cache flush problem.
    Hope this helps anyone who encountered the same problem...
    Thanks,
    Matthew

  • Buffer Cache Flush

    Hi All,
    What is the benefit of Flushing buffer cache. what are its effects and is it advisable to do it on production system.
    Thanks in Advance
    Edited by: Vikas Kohli on Jan 31, 2012 11:13 AM
    Edited by: Vikas Kohli on Jan 31, 2012 11:22 AM

    Asif,
    All blocks which are resides in the buffer cache get flushed. I think this needs correction. "However a buffer cannot become free if it is "pinned" (i.e. actualy in use) or if it is "dirty" (i.e. needs to be written to disc). There's nothing that Oracle can do about the pinned buffers, but it can write the content of the dirty buffers to disk before freeing them." Sir Jonathan @ flush buffer cache
    @ Vikas,
    In the above thread Sir Jonathan has written many most valuable inputs. Please check it.
    I think you will get more details from below link too : (Almost equal to Docs)
    http://asktom.oracle.com/pls/asktom/f?p=100:11:0::::P11_QUESTION_ID:7413988573867
    Flushing cache is generally and mostly used in Testing environment, untill and unless there is such need to do so.
    Flushing the buffer helps to find out more consistent results for sql traces.
    If you want to compare two different cases for performance, flush buffer cache and shared pool before execution of both of them.
    But this won't reduce LIO's (Logical I/O)
    Regards
    Girish Sharma

  • Solaris10 chown/chgrp on NFS mounts broken

    chown broken with a Solaris10 nfs server:
    nfs-server(Solaris10)# share -F nfs -o ro=.emp-eaw.ch,rw=hermes.emp-eaw.ch,root=hermes.emp-eaw.ch,anon=60001 -d "testings nfs" /nfstest
    nfs-client(Solaris10)# mount /nfstest10
    nfs-client(Solaris10)# mount|grep /nfstest
    /nfstest10 on uranos:/nfstest remote/read/write/setuid/devices/soft/actimeo=0/xattr/dev=4cca8d7 on Fri Mar 6 10:15:33 2009
    nfs-client(Solaris10)# mount|grep nfstest
    /nfstest8 on hercules:/nfstest remote/read/write/setuid/devices/soft/actimeo=0/xattr/dev=4cca8db on Fri Mar 6 10:22:37 2009
    nfs-client(Solaris10)# cd /nfstest10
    nfs-client(Solaris10)# touch xxx
    nfs-client(Solaris10)# ls -l xxx
    -rw-r--r-- 1 root root 0 Mar 6 10:16 xxx
    nfs-client(Solaris10)# chown guest xxx
    nfs-client(Solaris10)# ls -l xxx
    -rw-r--r-- 1 nobody root 0 Mar 6 10:16 xxx
    chown works with a Solaris 8 nfs server:
    nfs-server(Solaris8)# share -F nfs -o ro=.emp-eaw.ch,rw=hermes,root=hermes,anon=60001 -d "testings nfs" /nfstest
    nfs-client(Solaris8)# mount /nfstest8
    nfs-client(Solaris8)# nt|grep nfstest
    /nfstest8 on hercules:/nfstest remote/read/write/setuid/devices/soft/actimeo=0/xattr/dev=4cca8db on Fri Mar 6 10:22:37 2009
    nfs-client(Solaris8)# cd /nfstest8
    nfs-client(Solaris8)# touch xxx
    nfs-client(Solaris8)# ls -l xxx
    -rw-r--r-- 1 root root 0 Mar 6 10:22 xxx
    nfs-client(Solaris8)# chown guest xxx
    nfs-client(Solaris8)# ls -l xxx
    -rw-r--r-- 1 guest root 0 Mar 6 10:22 xxx
    any suggestions???
    kind regards
    G. Mohler, ICT/Eawag

    yes indeed!
    i added the option *"vers=3"* to the mountpoint in /etc/vfstab.
    now chown and chgrp work as well as with my Solaris-8 nfs server!
    thank you for your help
    kind regards
    :-) Gerard.

  • OAM - Identity Server Cache Flush to Access Server

    In OAM, a feature is that any user profile change in identity server could be updated to access server cache. Currently, we run into problem. the cache flush is a water fall process, from one identity server to an access server then to all other access servers. If one server hangs and does not respond, all servers may hang (that happens to us).
    I would like to know if you are using this feature and if you run into the same problem. You can also comment if this is a necessary feature and/or a troublesome feature. I would appreciate your any contribution. Thank you!
    Richard

    I am facing issue while setting up the automatic cache flush for access server.It looks like you were able to set up this. Can you please provide bit more details?. I have tried by setting -
    a. Parameter doAccessServerFlush in basedbparams as <NameValPair ParamName="doAccessServerFlush"
    Value="true"/>
    b. An Access Gate has been configured under /identity/AccessServerSDK
    My requirement is to flush the changes to the Group membership in a group. Are there any specific restrictions on what is included in user cache?.
    Thanks
    GK Bhat

  • Can't Disable "Allow Host Cache Flushing"

    I have problems if I have "Allow Host Cache Flushing" on when I digitize video. When I turn it off I have no problem whatsoever. The problem I have is that when I turn it off it won't stay. I'll press okay and then go and check it again and it is checked. Does anybody know what is going on.
    I had just brought a 1.5 hr program online and now all the video has dropped frames because it didn't stick.
    Dual 2.0 G5   Mac OS X (10.4.7)   FCP 5.0.4, XRaid Admin Software 1.5

    I have the same problem and it's not the cache batteries. It's just a preference corruption, whose solution I do not know. I doubt you are doing anything wrong. Any time I noticed a dropped from on capture, I go to the Raid Admin and re-do the performance settings, then it works fine for a long time.
    Steve Covello
    double wide post

  • After updating kernel and ZFS modules, system cannot boot

    Starting Import ZFS pools by cache file...
    [ 4.966034] VERIFY3(0 == zap_lookup(ddt->ddt_os, ddt->ddt_spa->spa_ddt_stat_object, name, sizeof (uint64_t), sizeof (ddt_histogram_t) / sizeof (uint64_t), &hht->ddt_histogram[type][class])) failed (0 == 6)
    [ 4.966100] PANIC at ddt.c:124:ddt_object_load()
    [*** ] A start job is running for Import ZFS pools by cache (Xmin Ys / no limit)
    And then occasionally I see
    [ 240.576219] Tainted: P O 3.19.2-1-ARCH #1
    Anyone else experiencing the same?

    Thanks!
    I did the same and it worked... kind of. The first three reboots it failed (but did not stop the system from booting) producing:
    zpool[426]: cannot import 'data': one or more devices is currently unavailable
    systemd[1]: zfs-import-cache.service: main process exited, code=exited, status=1/FAILURE
    The second boot also resulted in a kernel panic, but as far as I can tell unrelated to zfs.
    After reboot one and three imported the pool manually.
    From the fourth reboot on loading from cache file always succeeded. However, I it takes faily long (~8 seconds) and even shows
    [*** ] A start job is running for Import ZFS pools by cache (Xmin Ys / no limit)
    briefly. Altough I might only notice that because the recent updates sped up oder parts of the boot process. Did you observe a slowdown during boot time, too, kinghajj?
    Last edited by robm (2015-03-22 01:21:05)

  • Max File size in UFS and ZFS

    Hi,
    Any one can share what is max file size can be created in Solaris 10 UFS and ZFS ?
    What will be max size file compression using tar,gz ?
    Regards
    Siva

    from 'man ufs':
    A sparse file  can have  a  logical  size  of one terabyte.
    However, the  actual amount of data that can be stored
    in  a  file  is  approximately  one  percent  less  than one
    terabyte because of file system overhead.
    As for ZFS, well, its a 128bit filesystem, and the maximum size of a file or directory is 2 ^64^ bytes, which i think is somewhere around 8 exabyte (i.e 8192 petabyte), even though my calculator gave up on calculating it.
    http://www.sun.com/software/solaris/ds/zfs.jsp
    .7/M.
    Edited by: abrante on Feb 28, 2011 7:31 AM
    fixed layout and 2 ^64^

  • SunCluster, MPXIO, Clariion and ZFS?

    Hi,
    we have a 2 node cluster (SunCluster 3.2). Our Storage is a EMC Clariion CX700. We have created some zpools and integrated them into the suncluster.
    We cannot use PowerPath 5.1 and 5.2 for this because sun cluster and zfs is not supported in this environment. So, we want to use mpxio. Our question is, if there is a SP-Failover at the clariion, does mpxio support this and everything works fine without any problems?
    Thanks!
    Greets
    Björn

    Hi,
    What you need todo is the following.
    edit this file /kernel/drv/scsi_vhci.conf
    follow the directions of this link
    http://www.filibeto.org/sun/lib/nonsun/emc/SolarisHostConectivity.pdf?bcsi_scan_1BD4CB6F2E356E40=0&bcsi_scan_filename=SolarisHostConectivity.pdf
    regards
    Filip

Maybe you are looking for

  • I cannot turn up the volume during a phone call

    My "+" button on the side of my phone is broken and I cannot turn up the volume during a phone call. Is there any other way to control the volume during phone calls?

  • [svn] 4129: Fix for - the @inheritDoc tag in an ASDoc comment: is not pulling in a comment from UIComponent

    Revision: 4129 Author: [email protected] Date: 2008-11-17 14:57:15 -0800 (Mon, 17 Nov 2008) Log Message: Fix for - the @inheritDoc tag in an ASDoc comment: is not pulling in a comment from UIComponent QE Notes: None Doc Notes: None Bugs: SDK-18127, S

  • 850 Inbound - Newbie

    New to the whole EDI gateway so please bear with me. I've been tasked with getting the 850 inbound, 810 outbound and 856 outbound working. So far I've been able to get all of these to work with some semi-production data. We have a company that is wor

  • RDF + Virtual Private Database mechanism - r/w privileges to subject, ...

    Hi, is there a way how to give RDF subject/property/object read/write permissions with VPD policy? I only can give SELECT/UPDATE/DELETE permission to whole application table, but can I give read/write permissions to RDF subject, property or object? E

  • GDM no logging to /var/log complains

    I have a simple warning coming up: gdm-binary[pidnum]: WARNING: gdn_config_parse: Logdir /var/log/gdm does not exist or isn't a directory. Using ServAuthDir /var/lib/gdm Now if I create that directory and restart gdm everyone is happy. But since I us