NFS latency when Solaris 10 client mounts Linux NFS server(EMC NAS)

Hello,
One of our developers discovered a problem that for simplicity we call "latency". We have several 5.10 clients that we see the exact same symptoms on when NFS mounting our Celerra. The NAS is running a Linux variant "2.4.9-34.5406.EMC", but before you all jump on the "it's EMC's problem" bandwagon, let me explain. We set up an automated process (Perl) that watches an exported folder for the appearance of a request file (rand.req). When the request file comes in we rename the request file to (rand.sav) and then return a "report" named (rand.res). Very elegent I thought, and it runs at near lightspeed when only Linux NFS clients mount the share and create, monitor, delete, etc any files. In fact there is zero recorded latency from the time the report file appears and when the client detects it. But for all our Solaris 10 clients, they create the request file just fine, and the Perl process running on the Linux box sees the file instantaneously and returns the report, but it takes the Solaris client anywhere from 5 to up to 50 seconds before it see's any change in status for any files the Linux box manipulates. I've tried every possible combination of mount -o options there are including noac, rsize and wsize variants, vers=2, proto=udp, actimeo=0, etc, etc, etc. Nothing seems to be the magic bullet. nfsstat -c shows nothing out of the ordinary. There are no retransmits or dropepd packets anywhere in between, no firewall loads, no connectivity delays whatsoever. I'm completely out of ideas. Any ideas or clues would be greatly appreciated!
thanks
Dave

No specific recommendations. But maybe you can watch the cable and get more information.
Set up a case where the file has been created, then have the client check and snoop the cable at the same time. Does the client actually issue a directory check (or is it just displaying cached information)? Does the response contain the new file?
Something to test anyway...
Darren

Similar Messages

  • Oracle8i client on Linux, oracle8 server on Solaris

    Hi
    I have an Oracle8 server (8.0.6) on Solaris
    2.5.1 running on Sparc.
    I installed Oracle8i client (8.1.5) on linux
    (Redhat 6.1)
    With sqlplus it works, but using proc it fails.
    First I thought that it was my piece of program, but it fails with the demo software too, producing the same error message as my program.
    $ORACLE_HOME/precomp/demo/proc/ ansidyn1 and ansidyn2 fails when I try to issue a simple sql query with the following error:
    ORA-01002: fetch out of sequence
    The interesting thing is, that this error is generated when it tries to PREPARE the dynamic statement, and not when it tries to fetch.
    Any idea ???
    Gabor Dolla <[email protected]>
    null

    I have the same problem!I have already post it on this board!It seems like a bug!I think we should contact oracle corp. to solution it!

  • New files and folders on a Linux client mounting a Windows 2012 Server for NFS share do not inherit Owner and Group when SetGID bit set

    Problem statement
    When I mount a Windows NFS service file share using UUUA and set the Owner and Group, and set the SetGID bit on the parent folder in a hierarchy. New Files and folders inside and underneath the parent folder do not inherit the Owner and Group permissions
    of the parent.
    I am given to understand from this Microsoft KnowledgeBase article (http://support.microsoft.com/kb/951716/en-gb) the problem is due to the Windows implmentation of NFS Services not supporting the Solaris SystemV or BSD grpid "Semantics"
    However the article says the same functionality can acheived by using ACE Inheritance in conjunction with changing the Registry setting for "KeepInheritance" to enable Inheritance propagation of the Permissions by the Windows NFS Services.
    1. The Precise location of the "KeepInheritance" DWORD key appears to have "moved" in  Windows Server 2012 from a Services path to a Software path, is this documented somewhere? And after enabling it, (or creating it in the previous
    location) the feature seems non-functional. Is there a method to file a Bug with Microsoft for this Feature?
    2. All of the references on demonstrating how to set an ACE to achieve the same result "currently" either lead to broken links on Microsoft technical websites, or are not explicit they are vague or circumreferential. There are no plain Examples.
    Can an Example be provided?
    3. Is UUUA compatible with the method of setting ACE to acheive this result, or must the Linux client mount be "Mapped" using an Authentication source. And could that be with the new Flat File passwd and group files in c:\windows\system32\drivers\etc
    and is there an Example available.
    Scenario:
    Windows Server 2012 Standard
    File Server (Role)
    +- Server for NFS (Role) << -- installed
    General --
    Folder path: F:\Shares\raid-6-array
    Remote path: fs4:/raid-6-array
    Protocol: NFS
    Authentication --
    No server authentication
    +- No server authentication (AUTH_SYS)
    ++- Enable unmapped user access
    +++- Allow unmapped user access by UID/GID
    Share Permissions --
    Name: linux_nfs_client.host.edu
    Permissions: Read/Write
    Root Access: Allowed
    Encoding: ANSI
    NTFS Permissions --
    Type: Allow
    Principal: BUILTIN\Administrators
    Access: Full Control
    Applies to: This folder only
    Type: Allow
    Principal: NT AUTHORITY\SYSTEM
    Access: Full Control
    Applies to: This folder only
    -- John Willis, Facebook: John-Willis, Skype: john.willis7416

    I'm making some "major" progress on this problem.
    1. Apparently the "semantics" issue to honor SGID or grpid in NFS on the server side or the client side has been debated for some time. It also existed as of 2009 between Solaris nfs server and Linux nfs clients. The Linux community defaulted to declaring
    it a "Server" side issue to avoid "Race" conditions between simultaneous access users and the local file system daemons. The client would have to "check" for the SGID and reformulate its CREATE request to specify the Secondary group it would have to "notice"
    by which time it could have changed on the server. SUN declined to fix it.. even though there were reports it did not behave the same between nfs3 vs nfs4 daemons.. which might be because nfs4 servers have local ACL or ACE entries to process.. and a new local/nfs
    "inheritance" scheme to worry about honoring.. that could place it in conflict with remote access.. and push the responsibility "outwards" to the nfs client.. introducing a race condition, necessitating "locking" semantics.
    This article covers that discovery and no resolution - http://thr3ads.net/zfs-discuss/2009/10/569334-CR6894234-improved-sgid-directory-compatibility-with-non-Solaris-NFS-clients
    2. A much Older Microsoft Knowledge Based article had explicit examples of using Windows ACEs and Inheritance to "mitigate" the issue.. basically the nfs client "cannot" update an ACE to make it "Inheritable" [-but-] a Windows side Admin or Windows User
    [-can-] update or promote an existing ACE to "Inheritable"
    Here are the pertinent statements -
    "In Windows Services for UNIX 2.3, you can use the KeepInheritance registry value to set inheritable ACEs and to make sure that these ACEs apply to newly created files and folders on NFS shares."
    "Note About the Permissions That Are Set by NFS Clients
    The KeepInheritance option only applies ACEs that have inheritance enabled. Any permissions that are set by an NFS client will
    only apply to that file or folder, so the resulting ACEs created by an NFS client will
    not have inheritance set."
    "So
    If you want a folder's permissions to be inherited to new subfolders and files, you must set its permissions from the Windows NFS server because the permissions that are set by NFS clients only apply to the folder itself."
    http://support.microsoft.com/default.aspx?scid=kb;en-us;321049
    3. I have set up a Windows 2008r2 NFS server and mounted it with a Redhat Enteprise Linux 5 release 10 x86_64 server [Oct 31, 2013] and so far this does appear to be the case.
    4. In order to mount and then switch user to a non-root user to create subdirectories and files, I had to mount the NFS share (after enabling Anonymous AUTH_SYS mapping) this is not a good thing, but it was because I have been using UUUA - Unmapped Unix
    User Access Mapping, which makes no attempt to "map" a Unix UID/GID set by the NFS client to a Windows User account.
    To verify the Inheritance of additional ACEs on new subdirectories and files created by a non-root Unix user, on the Windows NFS server I used the right click properties, security tab context menu, then Advanced to list all the ACEs and looked at the far
    Column reflecting if it applied to [This folder only, or This folder and Subdirectories, or This folder and subdirectories and files]
    5. All new Subdirectories and files createdby the non-root user had a [Non-Inheritance] ACE created for them.
    6. I turned a [Non-Inheritance] ACE into an [Inheritance] ACE by selecting it then clicking [Edit] and using the Drop down to select [This folder, subdirs and files] then I went back to the NFS client and created more subdirs and files. Then back to the
    Windows NFS server and checked the new subdirs and folders and they did Inherit the Windows NFS server ACE! - However the UID/GID of the subdirs and folders remained unchanged, they did not reflect the new "Effective" ownership or group membership.
    7. I "believe" because I was using UUUA and working "behind" the UID/GID presentation layer for the NFS client, it did not update that presentation layer. It might do that "if" I were using a Mapping mechanism and mapped UID/GID to Windows User SIDs and
    Group SIDs. Windows 2008r2 no longer has a "simple" Mapping server, it does not accept flat text files and requires a Schema extension to Active Directory just to MAP a windows account to a UID/GID.. a lot of overhead. Windows Server 2012 accepts flat text
    files like /etc/passwd and /etc/group to perform this function and is next on my list of things to see if that will update the UID/GID based on the Windows ACE entries. Since the Local ACE take precedence "over" Inherited ACEs there could be a problem. The
    Inheritance appears to be intended [only] to retain Administrative rights over user created subdirs and files by adding an additional ACE at the time of creation.
    8. I did verify from the NFS client side in Linux that "Even though" the UID/GID seem to reflect the local non-root user should not have the ability to traverse or create new files, the "phantom" NFS Server ACEs are in place and do permit the function..
    reconciling the "view" with "reality" appears problematic, unless the User Mapping will update "effective" rights and ownership in the "view"
    -- John Willis, Facebook: John-Willis, Skype: john.willis7416

  • NFS4: Problem mounting NFS mount onto a Solaris 10 Client

    Hi,
    I am having problems mounting NFS mount point from a Linux-Server onto a Solaris 10 Client.
    In the following
    =My server IP ..*.120
    =Client IP ..*.100
    Commands run on Client:
    ==================
    # mount -o vers=3 -F nfs 172.25.30.120:/scratch/pvfs2 /scratch/pvfs2
    nfs mount: 172.25.30.120: : RPC: Rpcbind failure - RPC: Unable to receive
    nfs mount: retrying: /scratch/pvfs2
    nfs mount: 172.25.30.120: : RPC: Rpcbind failure - RPC: Unable to receive
    nfs mount: 172.25.30.120: : RPC: Rpcbind failure - RPC: Unable to receive
    # mount -o vers=4 -F nfs 172.25.30.120:/scratch/pvfs2 /scratch/pvfs2
    nfs mount: 172.25.30.120:/scratch/pvfs2: No such file or directory
    # rpcinfo -p
    program vers proto port service
    100000 4 tcp 111 rpcbind
    100000 3 tcp 111 rpcbind
    100000 2 tcp 111 rpcbind
    100000 4 udp 111 rpcbind
    100000 3 udp 111 rpcbind
    100000 2 udp 111 rpcbind
    1073741824 1 tcp 36084
    100024 1 udp 42835 status
    100024 1 tcp 36086 status
    100133 1 udp 42835
    100133 1 tcp 36086
    100001 2 udp 42836 rstatd
    100001 3 udp 42836 rstatd
    100001 4 udp 42836 rstatd
    100002 2 tcp 36087 rusersd
    100002 3 tcp 36087 rusersd
    100002 2 udp 42838 rusersd
    100002 3 udp 42838 rusersd
    100011 1 udp 42840 rquotad
    100021 1 udp 4045 nlockmgr
    100021 2 udp 4045 nlockmgr
    100021 3 udp 4045 nlockmgr
    100021 4 udp 4045 nlockmgr
    100021 1 tcp 4045 nlockmgr
    100021 2 tcp 4045 nlockmgr
    100021 3 tcp 4045 nlockmgr
    100021 4 tcp 4045 nlockmgr
    # showmount -e 172.25.30.120 (Server)
    showmount: 172.25.30.120: RPC: Rpcbind failure - RPC: Unable to receive
    Commands OnServer:
    ================
    program vers proto port
    100000 2 tcp 111 portmapper
    100000 2 udp 111 portmapper
    100021 1 tcp 49927 nlockmgr
    100021 3 tcp 49927 nlockmgr
    100021 4 tcp 49927 nlockmgr
    100021 1 udp 32772 nlockmgr
    100021 3 udp 32772 nlockmgr
    100021 4 udp 32772 nlockmgr
    100011 1 udp 796 rquotad
    100011 2 udp 796 rquotad
    100011 1 tcp 799 rquotad
    100011 2 tcp 799 rquotad
    100003 2 udp 2049 nfs
    100003 3 udp 2049 nfs
    100003 4 udp 2049 nfs
    100003 2 tcp 2049 nfs
    100003 3 tcp 2049 nfs
    100003 4 tcp 2049 nfs
    100005 1 udp 809 mountd
    100005 1 tcp 812 mountd
    100005 2 udp 809 mountd
    100005 2 tcp 812 mountd
    100005 3 udp 809 mountd
    100005 3 tcp 812 mountd
    100024 1 udp 854 status
    100024 1 tcp 857 status
    # showmount -e 172.25.30.120
    Export list for 172.25.30.120:
    /scratch/nfs 172.25.30.100,172.25.24.0/4
    /scratch/pvfs2 172.25.30.100,172.25.24.0/4
    Thank you, ~al

    I also tried to run Snoop on the client and wireshark on Server and following is what I see:
    One Server: Upon issuing mount command on client:
    # tshark -i eth1
    Running as user "root" and group "root". This could be dangerous.
    Capturing on eth1
    0.000000 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    0.205570 172.25.30.100 -> 172.25.30.120 Portmap V2 GETPORT Call MOUNT(100005) V:3 UDP
    0.205586 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    0.207863 172.25.30.100 -> 172.25.30.120 Portmap V2 GETPORT Call MOUNT(100005) V:3 UDP
    0.207869 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    2.005314 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    4.011005 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    5.206109 Dell_70:ad:29 -> SunMicro_70:ff:17 ARP Who has 172.25.30.100? Tell 172.25.30.120
    5.206277 SunMicro_70:ff:17 -> Dell_70:ad:29 ARP 172.25.30.100 is at 00:14:4f:70:ff:17
    5.216157 172.25.30.100 -> 172.25.30.120 Portmap V2 GETPORT Call MOUNT(100005) V:3 UDP
    5.216170 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    On Clinet Upon issuing mount command on client:
    # snoop -d bge1
    Using device /dev/bge1 (promiscuous mode)
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    atlas-pvfs2 -> pvfs2-io-0-3 PORTMAP C GETPORT prog=100005 (MOUNT) vers=3 proto=UDP
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (UDP port 111 unreachable)
    atlas-pvfs2 -> pvfs2-io-0-3 PORTMAP C GETPORT prog=100005 (MOUNT) vers=3 proto=UDP
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (UDP port 111 unreachable)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    pvfs2-io-0-3 -> * ARP C Who is 172.25.30.100, atlas-pvfs2 ?
    atlas-pvfs2 -> pvfs2-io-0-3 ARP R 172.25.30.100, atlas-pvfs2 is 0:14:4f:70:ff:17
    atlas-pvfs2 -> pvfs2-io-0-3 PORTMAP C GETPORT prog=100005 (MOUNT) vers=3 proto=UDP
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (UDP port 111 unreachable)
    Also I see the following on Client:
    # rpcinfo -p pvfs2-io-0-3
    rpcinfo: can't contact portmapper: RPC: Rpcbind failure - RPC: Failed (unspecified error)
    When I try the above rpcinfo command on Client and Server Snoop And wireshark(ethereal) outputs are as follows:
    Client # snoop -d bge1
    Using device /dev/bge1 (promiscuous mode)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    atlas-pvfs2 -> pvfs2-io-0-3 TCP D=111 S=872 Syn Seq=2065245538 Len=0 Win=49640 Options=<mss 1460,nop,wscale 0,nop,nop,sackOK>
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (TCP port 111 unreachable)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=2004 (Unknown), size = 48 bytes
    ? -> (multicast) ETHER Type=0003 (LLC/802.3), size = 90 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    pvfs2-io-0-3 -> * ARP C Who is 172.25.30.100, atlas-pvfs2 ?
    atlas-pvfs2 -> pvfs2-io-0-3 ARP R 172.25.30.100, atlas-pvfs2 is 0:14:4f:70:ff:17
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    atlas-pvfs2 -> pvfs2-io-0-3 TCP D=111 S=874 Syn Seq=2068043912 Len=0 Win=49640 Options=<mss 1460,nop,wscale 0,nop,nop,sackOK>
    pvfs2-io-0-3 -> atlas-pvfs2 ICMP Destination unreachable (TCP port 111 unreachable)
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> (multicast) ETHER Type=0000 (LLC/802.3), size = 52 bytes
    ? -> * ETHER Type=9000 (Loopback), size = 60 bytes
    Server # tshark -i eth1
    Running as user "root" and group "root". This could be dangerous.
    Capturing on eth1
    0.000000 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    0.313739 Cisco_3d:68:10 -> CDP/VTP/DTP/PAgP/UDLD CDP Device ID: MILEVA Port ID: GigabitEthernet1/0/16
    2.006422 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    3.483733 172.25.30.100 -> 172.25.30.120 TCP 865 > sunrpc [SYN] Seq=0 Win=49640 Len=0 MSS=1460 WS=0
    3.483752 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    4.009741 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    6.014524 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    6.551356 Cisco_3d:68:10 -> Cisco_3d:68:10 LOOP Reply
    8.019386 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    8.484344 Dell_70:ad:29 -> SunMicro_70:ff:17 ARP Who has 172.25.30.100? Tell 172.25.30.120
    8.484569 SunMicro_70:ff:17 -> Dell_70:ad:29 ARP 172.25.30.100 is at 00:14:4f:70:ff:17
    10.024411 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    12.030956 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    12.901333 Cisco_3d:68:10 -> CDP/VTP/DTP/PAgP/UDLD DTP Dynamic Trunking Protocol
    12.901421 Cisco_3d:68:10 -> CDP/VTP/DTP/PAgP/UDLD DTP Dynamic Trunking Protocol
    ^[[A 14.034193 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00  Cost = 0  Port = 0x8010
    15.691119 172.25.30.100 -> 172.25.30.120 TCP 866 > sunrpc [SYN] Seq=0 Win=49640 Len=0 MSS=1460 WS=0
    15.691138 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    16.038944 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    16.550760 Cisco_3d:68:10 -> Cisco_3d:68:10 LOOP Reply
    18.043886 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    20.050243 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    21.487689 172.25.30.100 -> 172.25.30.120 TCP 867 > sunrpc [SYN] Seq=0 Win=49640 Len=0 MSS=1460 WS=0
    21.487700 172.25.30.120 -> 172.25.30.100 ICMP Destination unreachable (Port unreachable)
    22.053784 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    24.058680 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    26.063406 Cisco_3d:68:10 -> Spanning-tree-(for-bridges)_00 STP Conf. Root = 32770/00:0a:b8:3d:68:00 Cost = 0 Port = 0x8010
    26.558307 Cisco_3d:68:10 -> Cisco_3d:68:10 LOOP Reply
    ~thank you for any help you can provide!!!

  • Solaris 9 NFS clients and Mac OS X 10.3.8 NFS Server issues

    I have a situation where I'm using a Mac OS X Server machine as my file server for heterogeneous mix of clients, Solaris 9 being one of them. The NFS server portion of Mac OS X seems to have some quirks. I made the move to a Mac OS X server from a combination of Linux and Solaris because it was touted as a good multi-platform server solution, but my NFS woes are souring my opinion of it
    I haven't been able to nail down the exact cause, but it seems that Sun's Gnome Destop 2.0 has problems starting up. The 'gconfd-2' process starts and never finishes what it's doing. I'm suspecting a problem with creating a lock file in the user's NFS mounted home directory. My workaround was to disable the Gnome Desktop 2.0 option, but it isn't very pleasant because many of my users liked it.
    Another strange issue that plagues the Solaris 9 and Linux (Fedora 2) users is that Mozilla, which is the primary email client for my users, complains that it "Could not initialize the browser's security component." It goes on to suggest that there may be a problem with read / write access with the user's profile directory or that there's no more room. Googling didn't turn up much about this problem, but the home directory share is nowhere near full and the permissions are such that eveyone can read and write to their own profile directory just fine. I've been able to work around this problem on Linux by removing the user's 'cert8.db' and 'key3.db' files before they run Mozilla, but this technique is failing to work on the Solaris 9 clients.
    All these problems seem to involve some strange file access issue, and the Linux and Solaris clients had no problems when I was using a Solaris box for home directory sharing via NFS, so it definitely seems like a problem with Mac OS X's NFS implementation.
    If anyone has come across this type of issue and has some information about a fix or a better workaround, I would love to hear from you. Thanks in advance!

    I just figured out today that the Gnome Desktop 2.0 problem is due to some part of Gnome not liking really long home directory paths. Mac OS X by default dictates that home directories be of the form /Network/Servers/(fqdn of file server)/Volumes/(volume name)(home directory share path), for example, /Network/Servers/xxx.myschool.edu/Volumes/Homes/userx. I haven't figured out what in Gnome doesn't like it, but it appears to be the gconf mechanism.
    The Mozilla strangeness still is happening though.
    Quote: schwenk wrote on Fri, 15 April 2005 11:09
    I haven't been able to nail down the exact cause, but it seems that Sun's Gnome Destop 2.0 has problems starting up. The 'gconfd-2' process starts and never finishes what it's doing. I'm suspecting a problem with creating a lock file in the user's NFS mounted home directory. My workaround was to disable the Gnome Desktop 2.0 option, but it isn't very pleasant because many of my users liked it.

  • Linux guy wants to install solaris 10 via nfs using centos as nfs server

    well in linux when installing this way I simply copy the dvd install image to an nfs share (or do mount -o loop "isoimage" /nfsmountdir), copy the boot.iso so a cdrom and do a "linux askmethod" during install and then specify the nfs server and dir.
    I want to learn sun and want to start by really understanding the install and how to set up slices etc. I would really like to do this using nfs using my centos box as the server as this method at least when installing linux is much mush faster.
    so can I do this and if so what are the commands. I did see some documentation but it was for installing from a sun nfs server. but I need to install from the centos as the server. I already have the iso's for the 5 cdroms as well as the dvd image. if possible I would obviously like to use the dvd image.
    thanks and remember I am a sun newbie (just bought my first solaris book this week so the more dummied down answer would be the best one for me). if it can be done could you give me the exact process.
    thanks a million (and one).

    The Sun network installer scripts only work on Solaris. So there's no instructions from Sun for configuring the daemons on Linux necessary to do the configuration.
    There are some guides on the net, but I don't know of one that tells the whole story.
    You might consider installing Solaris in a VM on the linux box, then configuring jumpstart on that. You could then see how the daemons on it are configured and replicate that configuration on the Linux machine.
    Darren

  • Parameters of NFS in Solaris 10 and Oracle Linux 6 with ZFS Storage 7420 in cluster without database

    Hello,
    I have ZFS 7420 in cluster and OS Solaris 10 and Oracle Linux 6 without DB and I need mount share NFS in this OS and I do not know which parameters are the best for this.
    Wich are the best parameters to mount share NFS in Solaris 10 or Oracle Linux 6?
    Thanks
    Best regards.

    Hi Pascal,
    My question is because when We mount share NFS in some servers for example Exadata Database Machine or Super Cluster  for best performance we need mount this shares with specific parameters, for example.
    Exadata
    192.168.36.200:/export/dbname/backup1 /zfssa/dbname/backup1 nfs rw,bg,hard,nointr,rsize=131072,wsize=1048576,tcp,nfsvers=3,timeo=600 0 0
    Super Cluster
    sscsn1-stor:/export/ssc-shares/share1      -       /export/share1     nfs     -       yes     rw,bg,hard,nointr,rsize=131072,wsize=131072,proto=tcp,vers=3
    Now,
    My network is 10GBE
    What happen with normal servers only with OS (Solaris and Linux)?
    Which parameters I need use for best performance?
    or are not necessary specific parameters.
    Thanks.
    Best regards.

  • NFS - Solaris 10 client from Ubuntu server gives Rpcbind error

    Hello All,
    New to Solaris, and I've been scouring the Internet to find a solution, but none have been produced. I'll start by giving you details about the setups, and then go into the error:
    Server Setup:
    Ubuntu 8.04
    Exports file ->
    /home/<folder> <Solaris 10 Server DNS name>(rw,no_subtree_check,async)
    Client Setup:
    Solaris 10
    Set /etc/default/nfs to have NFS_CLIENT_VERSMAX=3
    Ran svcadm -v enable -r network/nfs/client and then tried
    mount -F nfs <Ubuntu Server DNS name>:/home/<folder> /mnt/test/
    and all I ever get are Rpcbind failure - RPC: Timed Out and then it says it's retrying: /mnt/test
    I've gotten the firewall out of the way, I can ping the Ubuntu server from the Solaris server and vice versa, and I'm able to mount the Ubuntu NFS share on another Ubuntu machine perfectly, but I can't get it to mount on the Solaris server. If I specify v3 of NFS, that doesn't change anything. If I specify v4 of NFS, I get the error that the file or folder doesn't exist on the Ubuntu server.
    Any ideas? Any more info needed?

    This is the exact same problem I've been having. My server is Ubuntu 8.10, and the client is Solaris 10. This is on my home network, so I'm pretty confident it isn't a network issue. I do NFS all the time at work between Solaris machines, but I'm stumped on this one. I've noticed there are similar threads on the topic with no real answer that I have found --
    http://www.linuxquestions.org/questions/linux-networking-3/nfs-server-on-ubuntu-doesnt-play-nice-with-nfs-client-on-solaris-626508/
    I did a dfshares from the Solaris box, and I actually get a response listing the shares. Even though I can see it I still can't mount it. Here is what I see:
    bash-3.00# dfshares tabasco
    RESOURCE SERVER ACCESS TRANSPORT
    tabasco:/media/Shared tabasco - -
    bash-3.00# mount -F nfs -o ro tabasco:/media/Shared /mnt
    nfs mount: tabasco:/media/Shared: No such file or directory
    bash-3.00#
    NFS is working on the server, as I can mount it locally (see below)
    root@tabasco:/# cat /etc/exports
    /media/Shared *(ro,sync)
    root@tabasco:/# mount tabasco:/media/Shared /mnt
    root@tabasco:/# cd /mnt
    root@tabasco:/mnt# ls
    Videos lost+found Music Pictures Other
    root@tabasco:/mnt#
    Yes... my server''s name is tabasco... remember it's a home network... and I like Tabasco... :)

  • Turning off security to none when mounting nfs server

    My server has an option that allows me to set security to none when authenticating it's clients to mount its export filesystem. However OSX doesn't like this and requires some use of uid and gid to access a mounted drive. When I go to mount I receive an 'authentication error' from the terminal. I looked into the man pages for mount_nfs and it states:
    sec=<mechanism>
    +Force a specific security mechanism to be used for the mount, where mechanism is one of: krb5p, krb5i, krb5, or sys.+
    Authentication has to be one of these. Is there any way I can hack around this and set it up so this isn't a requirement on my machine. I want the equivalent of sec=none. My linux clients mount my NAS fine. This is the only thing stopping us from setting up our environment the way we desire. Does anyone have any suggestions? Thanks.
    Message was edited by: digitalapple

    rpc-svcgssd.service - RPC GSS-API Server Daemon
    Loaded: loaded (/usr/lib/systemd/system/rpc-svcgssd.service; disabled)
    Active: inactive (dead)
    Docs: man:rpc.svcgssd(8)
    rpc-gssd.service - RPC GSS-API Client Daemon
    Loaded: loaded (/usr/lib/systemd/system/rpc-gssd.service; disabled)
    Active: active (running) since Thu 2014-08-14 16:55:15 EDT; 17min ago
    Docs: man:rpc.gssd(8)
    Process: 253 ExecStart=/usr/bin/rpc.gssd $GSSD_OPTS (code=exited, status=0/SUCCESS)
    Main PID: 257 (rpc.gssd)
    CGroup: /system.slice/rpc-gssd.service
    └─257 /usr/bin/rpc.gssd
    Aug 14 16:55:14 home systemd[1]: Starting RPC GSS-API Client Daemon...
    Aug 14 16:55:15 home systemd[1]: Started RPC GSS-API Client Daemon.
    Aug 14 17:07:54 home rpc.gssd[637]: ERROR: Key table file '/etc/krb5.keytab' not found while beginning keytab scan for keytab 'FILE:/etc/krb5.keytab'
    Aug 14 17:07:54 home rpc.gssd[638]: ERROR: Key table file '/etc/krb5.keytab' not found while beginning keytab scan for keytab 'FILE:/etc/krb5.keytab'
    Aug 14 17:07:54 home rpc.gssd[638]: ERROR: Key table file '/etc/krb5.keytab' not found while beginning keytab scan for keytab 'FILE:/etc/krb5.keytab'

  • What do I need to install when I install oralce in linux and oracle client

    What do I need to install when I install oralce in linux and oracle client in windows?
    I want to connect oracle from windows to Linux, what do I need to install or setup? thanks.

    Are you trying to install the DB on Linux from your Windows Box? if so, you need to have an XWindows client on your client. You have to SSH or Telnet into the box, set your DISPLAY variable. In fact, it is all in your documentation for the DB install. We still don't understand what you are exactly asking.
    Do you have direct access to your Linux machine? If so, and it has a graphical GUI, then you can find install guides for most flavors of Linux at dizwell.com. His are most excellent and show examples, but do not go over the XWindows problem. You would have to resolve that first.
    A free XWindows environment for Windows is Cygwin.
    Once the DB is running on Linux then all you do is create a TNS entry on it on your client and it connects. It is the same as setting up a connection to connect to Oracle on a Windows box.

  • Solaris 10 NFS caching problems with custom NFS server

    I'm facing a very strange problem with a pure java standalone application providing NFS server v2 service. This same application, targeted for JVM 1.4.2 is running on different environment (see below) without any problem.
    On Solaris 10 we try any kind of mount parameters, system services up/down configuration, but cannot solve the problem.
    We're in big trouble 'cause the app is a mandatory component for a product to be in production stage in a while.
    Details follows
    System description
    Sunsparc U4 with SunOS 5.10, patch level: Generic_118833-33, 64bit
    List of active NFS services
    disabled   svc:/network/nfs/cbd:default
    disabled   svc:/network/nfs/mapid:default
    disabled   svc:/network/nfs/client:default
    disabled   svc:/network/nfs/server:default
    disabled   svc:/network/nfs/rquota:default
    online       svc:/network/nfs/status:default
    online       svc:/network/nfs/nlockmgr:default
    NFS mount params (from /etc/vfstab)
    localhost:/VDD_Server  - /users/vdd/mnt nfs - vers=2,proto=tcp,timeo=600,wsize=8192,rsize=8192,port=1579,noxattr,soft,intr,noac
    Anomaly description
    The server side of NFS is provided by a java standalone application enabled only for NFS v2 and tested on different environments like: MS Windows 2000, 2003, XP, Linux RedHat 10 32bit, Linux Debian 2.6.x 64bit, SunOS 5.9. The java application is distributed with a test program (java standalone application) to validate main installation and configuration.
    The test program simply reads a file from the NFS file-system exported by our main application (called VDD) and writes the same file with a different name on the VDD exported file-system. At end of test, the written file have different contents from the read one. Indeep investigation shows following behaviuor:
    _ The read phase behaves correctly on both server (VDD) and client (test app) sides, trasporting the file with correct contents.
    _ The write phase produces a zero filled file for 90% of resulting VDD file-system file but ending correctly with the same sequence of bytes as the original read file.
    _ Detailed write phase behaviour:
    1_ Test app wites first 512 bytes => VDD receive NFS command with offset 0, count 512 and correct bytes contents;
    2_ Test app writes next 512 bytes => VDD receive NFS command with offset 0, count 1024 and WRONG bytes contents: the first 512 bytes are zero filled (previous write) and last 512 bytes with correct bytes contents (current write).
    3_ Test app writes next 512 bytes => VDD receive NFS command with offset 0, count 1536 and WRONG bytes contents: the first 1024 bytes are zero filled (previous writes) and last 512 bytes with correct bytes contents (current write).
    4_ and so on...
    Further tests
    We tested our VDD application on the same Solaris 10 system but with our test application on another (Linux) machine, contacting VDD via Linux NFS client, and we don�t see the wrong behaviour: our test program performed ok and written file has same contents as read one.
    Has anyone faced a similar problem?
    We are Sun ISV partner: do you think we have enough info to open a bug request to SDN?
    Any suggestions?
    Many thanks in advance,
    Maurizio.

    I finally got it working. I think my problem was that I was coping and pasting the /etc/pam.conf from Gary's guide into the pam.conf file.
    There was unseen carriage returns mucking things up. So following a combination of the two docs worked. Starting with:
    http://web.singnet.com.sg/~garyttt/Configuring%20Solaris%20Native%20LDAP%20Client%20for%20Fedora%20Directory%20Server.htm
    Then following the steps at "Authentication Option #1: LDAP PAM configuration " from this doc:
    http://docs.lucidinteractive.ca/index.php/Solaris_LDAP_client_with_OpenLDAP_server
    for the pam.conf, got things working.
    Note: ensure that your user has the shadowAccount value set in the objectClass

  • NFS server unmount error w/ bind-mount

    Hello, please let me know if I should change the thread title.
    so I'm sharing a folder through NFS between two arch-linux pc's.
    Host1 is my desktop, Host2 is my laptop.
    I boot up my computer, rpcbind is started on boot-up
    I launch the server on Host2
    I launch the client on Host1
    on Host2, I do a bind-mount from some directory to the one I'm exporting
    I then unmount it, and it unmounts w/o error
    then I redo the bind-mount
    then, on the client, I mount the server's NFS share to some directory
    then, on the client, I unmount it w/o error
    then, on the server, I try to unmount the bind-mounted directory, but it won't, says it's "busy"
    if I restart the server, then I can unmount the bind-mount w/o error, but I believe it should work without me having to restart the server every time
    I want to bind a different directory to the exported directory.
    server
    $ systemctl start nfs-server
    client
    $ systemctl start nfs-client.target
    server
    $ mount test/ nfs/ --bind
    client
    $ mount Host1:/srv/nfs /mnt/nfs -t nfs
    $ umount /mnt/nfs
    server
    $ umount /srv/nfs
    umount: /srv/nfs: target is busy
    (In some cases useful info about processes that
    use the device is found by lsof(8) or fuser(1).)
    the output of
    $ lsof | grep /srv/nfs
    $ lsof | grep /srv
    did not show anything that had either the substring "nfs" or "rpc" in it...
    below is the status of both hosts rpcbind @startup (they were identical)
    $ systemctl status rpcbind
    rpcbind.service - RPC bind service
    Loaded: loaded (/usr/lib/systemd/system/rpcbind.service; static)
    Active: active (running) since Sat 2014-12-06 15:51:38 PST; 46s ago
    Process: 262 ExecStart=/usr/bin/rpcbind -w ${RPCBIND_ARGS} (code=exited, status=0/SUCCESS)
    Main PID: 264 (rpcbind)
    CGroup: /system.slice/rpcbind.service
    └─264 /usr/bin/rpcbind -w
    Dec 06 15:51:38 BabaLinux systemd[1]: Started RPC bind service.
    below is the status of the nfs-server @ startup then after it was launched
    $ systemctl status nfs-server
    nfs-server.service - NFS server and services
    Loaded: loaded (/usr/lib/systemd/system/nfs-server.service; disabled)
    Active: inactive (dead)
    $ systemctl status nfs-server
    nfs-server.service - NFS server and services
    Loaded: loaded (/usr/lib/systemd/system/nfs-server.service; disabled)
    Active: active (exited) since Sat 2014-12-06 15:53:44 PST; 21s ago
    Process: 483 ExecStart=/usr/sbin/rpc.nfsd $RPCNFSDARGS (code=exited, status=0/SUCCESS)
    Process: 480 ExecStartPre=/usr/sbin/exportfs -r (code=exited, status=0/SUCCESS)
    Main PID: 483 (code=exited, status=0/SUCCESS)
    CGroup: /system.slice/nfs-server.service
    Dec 06 15:53:44 BabaLinux exportfs[480]: exportfs: /etc/exports [1]: Neither 'subtree_check' or 'no_subtree_check' specified for export "NeoZelux:/srv/nfs".
    Dec 06 15:53:44 BabaLinux exportfs[480]: Assuming default behaviour ('no_subtree_check').
    Dec 06 15:53:44 BabaLinux exportfs[480]: NOTE: this default has changed since nfs-utils version 1.0.x
    Dec 06 15:53:44 BabaLinux systemd[1]: Started NFS server and services.
    below is the status of the nfs-client @ startup then after it was launched
    $ systemctl status nfs-client.target
    nfs-client.target - NFS client services
    Loaded: loaded (/usr/lib/systemd/system/nfs-client.target; disabled)
    Active: inactive (dead)
    $ systemctl status nfs-client.target
    nfs-client.target - NFS client services
    Loaded: loaded (/usr/lib/systemd/system/nfs-client.target; disabled)
    Active: active since Sat 2014-12-06 15:54:18 PST; 4s ago
    Dec 06 15:54:18 NeoZelux systemd[1]: Starting NFS client services.
    Dec 06 15:54:18 NeoZelux systemd[1]: Reached target NFS client services.

    NFSv4 operates in "namespaces" and you're basically messing with it's brain by bind-mounting (and unmounting) within the exported directory while it's live, without using the exportfs command (either via add/remove in /etc/exports or just putting it all on the commandline). You'll want to read up on section 7 of the RFC:
    https://tools.ietf.org/html/rfc5661
    In essence you're confusing it's brain -- when you add a bind mount, re-export and vice versa -- when removing it, deconfigure and re-export. It still might not work but you can try -- like Spider.007 I question your technique and methodology. It just feels like you're trying to do something... "wrong" here that should be designed in a better way.

  • NFS: server reboot - client problem with "file handle" - no unmont

    Hello,
    when I restart a NFS-server and a client has a NFS-share on it still mounted, that NFS-share becomes unusable on the client: everything causes the client complain about an old "file handle", even trying to unmount that share!
    So my question is: How to deal gracefully with a NFS-server reboot and avoiding this file handle problem?
    Apart from going to each client and unmount the NFS-share by hand?
    Thanks!

    Do I need to Give these values manually again as they are moved from Env to other?
    When we transport ID objects, the values stored in the channel get lost and have to be manually enetered......it happens even for DEV --> QA environment.
    Now your channel should reflect PROD-related data.
    2- Why my Receiver communiation channel coimg as
    Communication Channel | PRD_400 | RFC_QA3_Receiver:
    The channel name does not depend on the environment it is present....verify the name once again!
    Regards,
    Abhishek.

  • RMI Client on Linux issue.

    I am facing Java RMI problem on Suse Linux.
    RMI server running on solaris and RMI client running on Solaris ---> works fine
    RMI server running on solaris and RMI client running on Linux ---> gives error
    Pls note that RMI server running fine. When invoking remote method from RMI client on Suse Linux, I get following error.
    Java.rmi.UnmarshalException: error unmarshalling arguments; nested exception is:
    java.lang.ClassNotFoundException: RemoteVtskImpl_Stub
    Any help/pointers will be useful.

    Even we tried setting the CLASSPATH to point to those jar files. But still the problem persist.
    At run time is there any way to see what all the _stub files are downloaded in the JVM at the client machine (linux machine in our case) from the server machine (solaris machine in our case).                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • Novell Client for Linux causes Excel/Crossover to lock on file open

    When using Microsoft Office with Crossover, if I access a mounted NetWare
    drive using Novell Client for Linux it causes the office Open dialog to lock.
    The same action with an NCPFS mount works ok.
    Of course I understand if having Office/Crossover working is not exactly a
    priority but my company will need it if we are to migrate to NLD

    Malt,
    It appears that in the past few days you have not received a response to your
    posting. That concerns us, and has triggered this automated reply.
    Has your problem been resolved? If not, you might try one of the following options:
    - Do a search of our knowledgebase at http://support.novell.com/search/kb_index.jsp
    - Check all of the other support tools and options available at
    http://support.novell.com.
    - You could also try posting your message again. Make sure it is posted in the
    correct newsgroup. (http://support.novell.com/forums)
    Be sure to read the forum FAQ about what to expect in the way of responses:
    http://support.novell.com/forums/faq_general.html
    If this is a reply to a duplicate posting, please ignore and accept our apologies
    and rest assured we will issue a stern reprimand to our posting bot.
    Good luck!
    Your Novell Product Support Forums Team
    http://support.novell.com/forums/

Maybe you are looking for