Verifying NFS

Hello all,
Hopefully someone can help me out as I dont know linux at all to verify some information.
I have MARS configured to backup to a NFS server. I recently received a notification from MARS stating that one of the backups failed.
Because I do not have access to the NFS server (Managed Services is why), what other ways can I verify that the NFS backup is succeeding? Whenever I attempt to use the GUI to pull raw messages from the NFS server it essentially times out.
Anything on the cli available? Any and all help/ideas are appreciated.
MARS is running 4.2.6 ( 2458 ). Yea, old, but customer absolutely refuses to give us the time of day to upgrade it.

Also there is a 'pnexp' command available on the CLI. Even tough it is meant to export 4.x database before importing it into a new 5.x box (Generation 1). It can be used for 'immediate backups' as per the Cisco engineer on this thread:
http://forum.cisco.com/eforum/servlet/NetProf?page=netprof&forum=Expert%20Archive&topic=Security&topicID=.ee7f99a&CommCmd=MB%3Fcmd%3Dpass_through%26location%3Doutline%40%5E1%40%40.2cbeaa98/18#selected_message
Once you enter the pnexp CLI
Do the following:
export config 10.1.1.1:/archive.
Then enter the following command:
pnexp> status
Data exporting process is currently running, please use command 'log {all|recent}' to view running logs and/or progress.
pnexp> log
Aug 25 09:44:47.192 2008@LM_INFO@Thread 1024:Number of events exported: 1227522, ~0.18% completed, overall speed = 30688.05 eps
Aug 25 09:44:47.192 2008@LM_INFO@Thread 1024:Estimated-Time-To-Complete: 6 hours 18 minutes
CTRL-C
As you can see I'm doing a full backup here (config+event/report data etc), don't run that just to 'test' NFS. Its better to use the 'config' option :)
Regards
Farrukh

Similar Messages

  • OS X Mavericks: Connecting to ZFS on Ubuntu Through NFS

    I have a working zpool and dataset set on my home server. I want to setup NFS to access ZFS over my home network on my Mac. The server is running Ubuntu 14.04; my Mac is running OS 10.9.2 Mavericks.
    On the Ubuntu side, I have...
    # apt-get install nfs-kernel-server
    # zfs set [email protected]/24,insecure tank/documents
    should allow 192.168.0.0 - 192.168.0.255
    insecure is needed so the mac connects on ports bigger than 1024
    # zfs share tank/documents
    # service nfs-kernel-server startcommented out && grep -q '^[[:space:]]*[^#]*/' $export_files in /etc/init.d/nfs-kernel-server because I can't start the server with an empty /etc/exports/ file
    # showmount -e to verify directory is mounted
    The output of showmount -e is this.
    Export list for ubuntu1404:
    /tank/documents *
    On the Mac side, I have...
    $ rpcinfo -p 192.168.0.3 to verify NFS is visible on the network
    $ mount -t nfs 192.168.0.3:/tank/documents /Users/me/Remote
    This last step is where I'm stuck. I keep getting "Operation not permitted".
    I've been reading around, but I've only found old tutorials that use NFS Mounts from Disk Utility. Does anyone have any up-to-date tutorials?

    Mae8se wrote:
    Okay, so I must have the syntax for mount wrong because I can connect through Finder's Connect to server menu item.
    nfs://192.168.0.3/tank/documents
    Not necessarily. NFS is ancient UNIX technology. It was designed for a world where only root mounted volumes. If you mount through the Finder, then root is doing the mounting for you. All bets are off if you try from the command line.
    What is Autofs?
    Here is the documentation: http://images.apple.com/business/docs/Autofs.pdf
    Autofs was designed with NFS in mind, so it should work well with it. It will be like having the Finder do the mount for you whenever you attempt to access the mount point. In theory, it should unmount the server too when you are no longer using it.

  • How do I verify Direct NFS is being used by RMAN

    I have a two node 11.2.0.2 RAC database running on AIX 6.1 TL6. We using NetApp storage for NFS. Direct NFS is configured using the /etc/oranfstab entries and the ODM libraries are loaded on each instance "Oracle instance running with ODM: Oracle Direct NFS ODM Library Version 3.0". Here are the contents of /etc/filesystems
    /export:
    dev = /vol/volnshr001_oraexport/oraexport
    vfs = nfs
    nodename = filer1
    mount = true
    options = rw,bg,hard,rsize=32768,wsize=32768,timeo=600,vers=3,proto=tcp,noac
    account = false
    /archivelogs:
    dev = /vol/volnshr001_oraclearchive
    vfs = nfs
    nodename = filer1
    mount = true
    options = rw,bg,hard,rsize=32768,wsize=32768,timeo=600,vers=3,proto=tcp,noac
    account = false
    /archivelogs2:
    dev = /vol/volnshr002_oraclearchive
    vfs = nfs
    nodename = filer2
    mount = true
    options = rw,bg,hard,rsize=32768,wsize=32768,timeo=600,vers=3,proto=tcp,noac
    account = false
    The oranfstab has the following entries:
    server: filer1
    path: 192.168.2.22
    export: /vol/volnshr001_oraclearchive mount:/archivelogs
    server: filer2
    path: 192.168.2.23
    export: /vol/volnshr002_oraclearchive mount:/archivelogs2
    server: filer1
    path: 192.168.2.22
    export: /vol/volnshr001_oraexport/oraexport mount:/export
    I tested if dNFS is being used by creating a tablespace on the NFS volumes and then querying v$dnfs_servers.
    SQL> create tablespace dnfs_ts datafile '/archivelogs/tiboratst/tibuat/dnfs_ts_01.dbf' size 100M;
    Tablespace created.
    SQL> !
    oracle@tibora30t[tibuat1]-/u01/oracle/11.2.0/dbhome_1/lib >ls -ltr /archivelogs/tiboratst/tibuat
    total 24811120
    -rw-r----- 1 oracle asmadmin 2896360960 Aug 18 17:59 1fmkanau_1_2
    -rw-r----- 1 oracle asmadmin 2896360960 Aug 18 17:59 1fmkanau_1_1
    -rw-r----- 1 oracle asmadmin 3377919488 Aug 18 17:59 1emkanau_1_2
    -rw-r----- 1 oracle asmadmin 3377919488 Aug 18 17:59 1emkanau_1_1
    -rw-r----- 1 oracle asmadmin 104865792 Aug 18 18:01 dnfs_ts_01.dbf
    oracle@tibora30t[tibuat1]-/u01/oracle/11.2.0/dbhome_1/lib >exit
    SQL> select * from v$dnfs_servers;
    ID
    SVRNAME
    DIRNAME
    MNTPORT NFSPORT WTMAX RTMAX
    1
    filer1
    /vol/volnshr001_oraclearchive
    4046 2049 65536 65536
    The alert logs also displayed the following messages:
    Direct NFS: channel id [0] path [192.168.2.22] to filer [filer1] via local [] is UP
    Direct NFS: channel id [1] path [192.168.2.22] to filer [filer1] via local [] is UP log also displayed the following messages:
    However, when I try to do an RMAN backup I'm not seeing any messages which indicate whether or not dNFS is being used. I tried to to a backup using the second mount (/archivelogs2) but the corresponding server (filer2) was not visible in v$dnfs_servers.
    Is there any special configuration necessary for RMAN backups to use dNFS?
    Thanks,
    Leighton

    Tiger install will take up 12gb if you don't go through and do a custom install and remove stuff that you don't need.
    There is a program that you can download that tells you the size of each folder, get it here: http://www.macupdate.com/info.php/id/13006

  • 10.6.2 failed to install due to 'nfs error', audio no longer works

    macbook pro, all other updates have installed just fine, except for 10.6.2
    I get this in the log:
    11/10/09 3:58:15 PM com.apple.kextcache410 Created mkext archive //System/Library/Caches/com.apple.kext.caches/Startup/Extensions.mkext.
    11/10/09 3:58:15 PM com.apple.kextcache403 Failed to mount helper partition: error 0XF8DA000C (DA err# 0x0c).
    11/10/09 3:58:15 PM com.apple.kextd10 kextcache error while updating / (error count: 4)
    11/10/09 3:58:15 PM com.apple.kextcache403 /: trouble updating one or more helper partitions.
    11/10/09 3:58:15 PM com.apple.kextd10 async child pid 403 exited with status 70
    11/10/09 3:58:15 PM com.apple.kextd10 kextcache error while updating / (error count: 5)
    11/10/09 3:58:15 PM installd376 PackageKit: Install Failed: (null)
    Error Domain=PKInstallErrorDomain Code=121 UserInfo=0x104d22740 "An error occurred while updating system extension information." Underlying Error=(Error Domain=NSPOSIXErrorDomain Code=70 "The operation couldn’t be completed. Stale NFS file handle") {
    NSLocalizedDescription = "An error occurred while updating system extension information.";
    NSUnderlyingError = "Error Domain=NSPOSIXErrorDomain Code=70 \"The operation couldn\U2019t be completed. Stale NFS file handle\"";
    arguments = (
    "-update-volume",
    nothing is mounted (i rebooted before applying). not sure why it's giving that NFS error. it's consistent tho.. if i reboot and try the manual install of the patch, I get the same error.
    after the failed patch "installed" it left my sound not working... so far everything else appears to work. any thoughts on fixing this are appreciated. i ran diskutil and 'fixed perms' and 'verified disk' ... patch still fails.

    Try reinstalling the patch. If you get all the way through the installation, then see if your sound works. If not, then I suggest you do a repair install of Sno by installing Sno over Sno as in an upgrade install. Such an install, while usually very safe, still is an install so be appropriated backed up if you proceed.

  • How to mount NFS into a zone or how to use the same IP in every zone?

    Hi,
    we are using a NFS server in a different subnet (= different network interface; nxge1) than our 'normal' network (= nxge0).
    Now I want to mount serveral NFS directories into our zones.
    First I thought the follwoing is possible:
    In the global zone I just mount the filesystem into its place:
    mount 10.11.3.1:/vol/vol_nfs_zones/zoneNNN_mnt /zones/zoneNNN/root/mnt
    When the zone is halted this mount works, but when I start the zone, I receive the error:
    # zoneadm -z zoneNNN boot
    zoneadm: zone 'zoneNNN': These file-systems are mounted on subdirectories of /zones/zoneNNN/root:
    zoneadm: zone 'zoneNNN': /zones/zoneNNN/root/mnt
    zoneadm: zone 'zoneNNN': call to zoneadmd failed
    When zoneNNN is already running and I try to mount it, I receive
    nfs mount: mount: /zones/zoneNNN/root/mnt: Device busy
    So the question is: How do I mount something in the gloabl zone from a NFS server into a zone?
    The only alternative I see is to make the interface nxge1 available in every zone and give each an own IP in our storage network... But sadly this isn't an alternative because we don't have so many free IPs in that net! (Or is there a trick to use the same IP in the global and non-global zones for only this interface??)
    Does anyone have a solution for our problem? I cannot believe that SUN misses this!
    Any hints are more than welcome!

    Mounting it by NFS in the global zone and trying to add this by the "add fs" and "type=lofs"-option yields to an error, when installing the zone:
    cannot verify fs /data: NFS mounted file-system.
    A local file-system must be used.
    zoneadm: zone zoneNNN failed to verifyOur storage network is not accessable from other nets. (And I think it is a very bad idea to route storage network traffic...)
    So I think we def. have to increase our storage network. :-(

  • No longer mount NFS volumes with 2010.q1 release:

    Has anyone else seen this? I can no longer mount any of my NFS volumes with root squash since upgrading to 2010.q1. I have a ticket open with Sun (72644622), but everyone is apparently off/away.
    Before they all went off, Sun Eng. suggested setting max supported NFS version to 3, but this of course did not work. Hosts trying to mount the volumes error out with:
    call_verify: server HOST_IP_HERE requires stronger authentication.
    I have double checked that the atuh is set to sys (not krb), but it looks like the nfsd is not respecting this.
    Of course this has knocked all our VMware mounts offline as well, and using krb for these is out of the question....
    Had to update to support the new FC card...which works great...but me needs me nfs shares....
    Fishworks doods, stop yelling at drives and help a brother out ;)

    I wish I could tell you, but it seems we lost a CIM card. Worse yet, this failure was completely SILENT. Not a single "problem" listed in the logs. However, when attaching another J4400, I noticed an alarm LED on the failed CIM and verified that the storage was only seen by 1 path. Kind of disconcerting that availability could be reduced significantly without so much as an email.
    I'll let you know if we can get to the new release later this week.
    Charles

  • Custom NFS share point directory showing up on all network machines

    Hi,
    I'm in the process of migrating our 10.4 PowerMac server to a Mac Pro (running 10.5). I've been trying to recreate our 10.4 server setup as much as possible and so far I've only come across one annoying issue.
    We have fink installed on the server and under our 10.4 setup the /sw directory was set up as an NFS automounted share point with a custom mount point of '/sw'. I.e. users logging into client machines saw a /sw directory and could work with that. This made it easier to add fink packages as I only needed to do this on one machine (the server). This setup worked very well under 10.4 and had been working stably for the last couple of years.
    As we now have (for another month or two at least) a mix of intel and Power PC machines, I don't want to share out the (intel) server version of fink to all clients. In Server Admin, I have chosen to set the NFS protocol options to specify the IP address of just one client (an intel machine). I am only using NFS to share this directory. The plan is to add more client IP addresses as we get more intel machines.
    This works for the one intel client machine. Logging in via the GUI or via ssh allows you to run programs located under the /sw directory. The problem is that a phantom /sw directory appears on all client machines, even though their IP addresses are not specified in Server Admin. The /sw directory has root/wheel permissions (for user/group) and attempting to list its contents returns 'Operation not permitted' (even with 'sudo ls /sw').
    If I use Directory Utility to remove the connection to the Directory server on our main server, then the /sw directory becomes owned by root/admin and I can remove it (it appears empty). Reconnecting to the Directory Server changes the permissions back to root/wheel. It is also worth noting that when I first installed fink on the server (in /sw) the act of making this a share point also changed the permissions on /sw to root/wheel meaning that I couldn't access the fink programs that I had only just installed (this forced me to reinstall fink in /Volumes/Data/fink).
    Has anyone else noticed this behavior? It almost seems like Server Admin is not honoring the list of IP addresses that are being listed as targets for client machines. I had planned to install fink locally on the PowerPC clients until we upgrade them to intel machines. However, I would then also have to install fink somewhere other than /sw as I can't write to that directory. I would presume that this behavior should happen on any NFS share point that is trying to automount to a custom mount point on a client. Can anyone else verify this?
    Regards,
    Keith

    As a footnote. I have now removed my shared fink installation. It is no longer listed as an NFS sharepoint in Server Admin and running the 'showmount -e' command does not list it. However, a /sw directory is still being created on the server and on the client machines on our network.
    This is perplexing and frustrating. There is no sharepoint any more. I rebooted the server but it makes no difference. I removed the /sw directory (on the server) by booting the machine in target firewire mode and removing it by using a 2nd machine. But following the restart, it appeared again.
    This suggests that once you make any custom mountpoint (using NFS sharing) then you will forever be stuck with a directory at the root level of all your clients which you can not remove.
    Keith

  • New files and folders on a Linux client mounting a Windows 2012 Server for NFS share do not inherit Owner and Group when SetGID bit set

    Problem statement
    When I mount a Windows NFS service file share using UUUA and set the Owner and Group, and set the SetGID bit on the parent folder in a hierarchy. New Files and folders inside and underneath the parent folder do not inherit the Owner and Group permissions
    of the parent.
    I am given to understand from this Microsoft KnowledgeBase article (http://support.microsoft.com/kb/951716/en-gb) the problem is due to the Windows implmentation of NFS Services not supporting the Solaris SystemV or BSD grpid "Semantics"
    However the article says the same functionality can acheived by using ACE Inheritance in conjunction with changing the Registry setting for "KeepInheritance" to enable Inheritance propagation of the Permissions by the Windows NFS Services.
    1. The Precise location of the "KeepInheritance" DWORD key appears to have "moved" in  Windows Server 2012 from a Services path to a Software path, is this documented somewhere? And after enabling it, (or creating it in the previous
    location) the feature seems non-functional. Is there a method to file a Bug with Microsoft for this Feature?
    2. All of the references on demonstrating how to set an ACE to achieve the same result "currently" either lead to broken links on Microsoft technical websites, or are not explicit they are vague or circumreferential. There are no plain Examples.
    Can an Example be provided?
    3. Is UUUA compatible with the method of setting ACE to acheive this result, or must the Linux client mount be "Mapped" using an Authentication source. And could that be with the new Flat File passwd and group files in c:\windows\system32\drivers\etc
    and is there an Example available.
    Scenario:
    Windows Server 2012 Standard
    File Server (Role)
    +- Server for NFS (Role) << -- installed
    General --
    Folder path: F:\Shares\raid-6-array
    Remote path: fs4:/raid-6-array
    Protocol: NFS
    Authentication --
    No server authentication
    +- No server authentication (AUTH_SYS)
    ++- Enable unmapped user access
    +++- Allow unmapped user access by UID/GID
    Share Permissions --
    Name: linux_nfs_client.host.edu
    Permissions: Read/Write
    Root Access: Allowed
    Encoding: ANSI
    NTFS Permissions --
    Type: Allow
    Principal: BUILTIN\Administrators
    Access: Full Control
    Applies to: This folder only
    Type: Allow
    Principal: NT AUTHORITY\SYSTEM
    Access: Full Control
    Applies to: This folder only
    -- John Willis, Facebook: John-Willis, Skype: john.willis7416

    I'm making some "major" progress on this problem.
    1. Apparently the "semantics" issue to honor SGID or grpid in NFS on the server side or the client side has been debated for some time. It also existed as of 2009 between Solaris nfs server and Linux nfs clients. The Linux community defaulted to declaring
    it a "Server" side issue to avoid "Race" conditions between simultaneous access users and the local file system daemons. The client would have to "check" for the SGID and reformulate its CREATE request to specify the Secondary group it would have to "notice"
    by which time it could have changed on the server. SUN declined to fix it.. even though there were reports it did not behave the same between nfs3 vs nfs4 daemons.. which might be because nfs4 servers have local ACL or ACE entries to process.. and a new local/nfs
    "inheritance" scheme to worry about honoring.. that could place it in conflict with remote access.. and push the responsibility "outwards" to the nfs client.. introducing a race condition, necessitating "locking" semantics.
    This article covers that discovery and no resolution - http://thr3ads.net/zfs-discuss/2009/10/569334-CR6894234-improved-sgid-directory-compatibility-with-non-Solaris-NFS-clients
    2. A much Older Microsoft Knowledge Based article had explicit examples of using Windows ACEs and Inheritance to "mitigate" the issue.. basically the nfs client "cannot" update an ACE to make it "Inheritable" [-but-] a Windows side Admin or Windows User
    [-can-] update or promote an existing ACE to "Inheritable"
    Here are the pertinent statements -
    "In Windows Services for UNIX 2.3, you can use the KeepInheritance registry value to set inheritable ACEs and to make sure that these ACEs apply to newly created files and folders on NFS shares."
    "Note About the Permissions That Are Set by NFS Clients
    The KeepInheritance option only applies ACEs that have inheritance enabled. Any permissions that are set by an NFS client will
    only apply to that file or folder, so the resulting ACEs created by an NFS client will
    not have inheritance set."
    "So
    If you want a folder's permissions to be inherited to new subfolders and files, you must set its permissions from the Windows NFS server because the permissions that are set by NFS clients only apply to the folder itself."
    http://support.microsoft.com/default.aspx?scid=kb;en-us;321049
    3. I have set up a Windows 2008r2 NFS server and mounted it with a Redhat Enteprise Linux 5 release 10 x86_64 server [Oct 31, 2013] and so far this does appear to be the case.
    4. In order to mount and then switch user to a non-root user to create subdirectories and files, I had to mount the NFS share (after enabling Anonymous AUTH_SYS mapping) this is not a good thing, but it was because I have been using UUUA - Unmapped Unix
    User Access Mapping, which makes no attempt to "map" a Unix UID/GID set by the NFS client to a Windows User account.
    To verify the Inheritance of additional ACEs on new subdirectories and files created by a non-root Unix user, on the Windows NFS server I used the right click properties, security tab context menu, then Advanced to list all the ACEs and looked at the far
    Column reflecting if it applied to [This folder only, or This folder and Subdirectories, or This folder and subdirectories and files]
    5. All new Subdirectories and files createdby the non-root user had a [Non-Inheritance] ACE created for them.
    6. I turned a [Non-Inheritance] ACE into an [Inheritance] ACE by selecting it then clicking [Edit] and using the Drop down to select [This folder, subdirs and files] then I went back to the NFS client and created more subdirs and files. Then back to the
    Windows NFS server and checked the new subdirs and folders and they did Inherit the Windows NFS server ACE! - However the UID/GID of the subdirs and folders remained unchanged, they did not reflect the new "Effective" ownership or group membership.
    7. I "believe" because I was using UUUA and working "behind" the UID/GID presentation layer for the NFS client, it did not update that presentation layer. It might do that "if" I were using a Mapping mechanism and mapped UID/GID to Windows User SIDs and
    Group SIDs. Windows 2008r2 no longer has a "simple" Mapping server, it does not accept flat text files and requires a Schema extension to Active Directory just to MAP a windows account to a UID/GID.. a lot of overhead. Windows Server 2012 accepts flat text
    files like /etc/passwd and /etc/group to perform this function and is next on my list of things to see if that will update the UID/GID based on the Windows ACE entries. Since the Local ACE take precedence "over" Inherited ACEs there could be a problem. The
    Inheritance appears to be intended [only] to retain Administrative rights over user created subdirs and files by adding an additional ACE at the time of creation.
    8. I did verify from the NFS client side in Linux that "Even though" the UID/GID seem to reflect the local non-root user should not have the ability to traverse or create new files, the "phantom" NFS Server ACEs are in place and do permit the function..
    reconciling the "view" with "reality" appears problematic, unless the User Mapping will update "effective" rights and ownership in the "view"
    -- John Willis, Facebook: John-Willis, Skype: john.willis7416

  • Word 2008 for Mac and NFS mounted home directories "Save File" issues

    Greetings everyone,
    (Long time lurker, first time poster here)
    I admin a small network (under 20 workstaitons) with a centralized NFS server, with user home directories mounted via NFS upon login.  Users are authenticated via LDAP.  This is all working fine, there is no problem here.  The problem lies when my users use Microsoft Word 2008 for Mac.  When they attempt to save a file to thier Desktop (or Documents or any folder under thier home dir) they are met with the following message:
    (dialog box popup)
    "Word cannot save or create this file.  The disk maybe be full or write-protected.  Try one or more of the following: * Free more memory. * Make sure the disk you want to save the file on is not full, write-protected or damaged. (document-name.ext)"
    This happens regardless of file format (Doc, Docx, Txt) and regardless of saved location under the network mounted dir.  I've noticed that when saving Word creates a .tmp file in the target directory, which only further confuses me to the underlying cause of the issue.
    When users logon to a local machine account and attempt the save, there is no issue.
    I have found many posts in other commuity forums, including this one, indicating that the issue is a .TempoaryItems folder in the root of the mounted directory.  This folder already exists and is populated with entries such as "folder.2112" (where 2112 is the uid of the LDAP user).  I find other posts indicating that this is an issue with Word:2008 and OSX10.8, with finger pointing in either direction, but no real solution.
    I have installed all Office for Mac updates from Microsoft (latest version 12.3.6).
    I have verified permissions of the user's home dir.
    I have also ensured that this issue effects ONLY Microsoft Office 2008 for Mac apps, LibreOffice and other applications have no issue.
    Does *ANYONE* have a solution or workaround for this issue?  While we're trying to phase Microsoft products out, getting users to ditch Word and Excel is difficult without removing them from systems completely.  So any pointers or help would be greatly appreciated.
    Thanks.
    ~k

    I can't tell you how to fix bugs in an obsolete version of Office, but a possible workaround is to use mobile home directories under OS X Server. The home directories are hosted locally and synced with the server.

  • Files saved to NFS share by MS word cause finder to loose sync w/filesystem

    I am posting this bug report(#4699323, opened Aug 24th, 2006) I have open with Apple because there is no indication that they have even noticed it. Please reply if you have similar problems, questions, or suggestions. Bug report follows.
    24-Aug-2006 04:17 PM Erik Meitner:
    Summary:
    Files saved to an NFS volume by Microsoft Word dissappear from the Finder. Many times "Ghosted" icons of "Word Work Files" are left behind.
    Steps to Reproduce & Actual Results:
    1. Mount an NFS volume from a server(Mac or Linux, doesn't matter. NFS options do not matter.) from the Finder or command line.
    2. Open a Finder window to the mount point.
    3. Create a word document(test.doc), save to the NFS mount point.
    4. Make sure both the Word window and Finder window are visible.
    5. Type a single letter, Save.
    6. "Word Work File L_200098113"(the number varies.) apears and test.doc dissapears from the folder. Then the icon for "Word Work File L_200098113" goes "ghosted" (test.dc IS still there, just not shown. You can verify this in the shell )
    7. Type a letter, Save.
    8. The work file dissapears and is replaces by "Word Work File L_200098113"
    9. Rinse, Repeat ad infinitum.
    Expected Results:
    File saved & Finder does NOT get out of sync with the filesystem.
    Regression:
    * Only happens on NFS volumes.
    * Only happens with MS Word.
    * It happens on Panther(10.3.9) and Tiger(up to 10.4.7)
    * "Fast saves" in Word has no effect.
    * The user has full access to the root of the NFS volume.(a .TemporaryItems folder IS being created)
    * Verified that problem happens with all Versions of Office 2004 from 11.2.3 to 11.2.6(current).
    * Happens with ALL users: admin, local, and non-local(LDAP authenticated, network homes).
    Notes:
    * The work files(with "ghosted" icons) never show up when listing the folder via the CLI.
    * The original(test.doc) always shows in the CLI.
    * Relaunching the Finder usually causes the "missing" files to show again, but not always.
    * Problem is exacerbated by users losing faith in the system, thus saving very frequently, guaranteeing that the problem will always happen.
    * Only workaround is to have users work on files "off the network" on a local disk.
    30-Aug-2006 01:45 PM Erik Meitner:
    Screen shot attached showing initial state of Finder.
    '1.tiff' was successfully uploaded
    30-Aug-2006 01:46 PM Erik Meitner:
    Screen shot attached showing state of Finder after 1st save.
    '2.tiff' was successfully uploaded
    30-Aug-2006 01:47 PM Erik Meitner:
    Screen shot attached showing state of Finder after 3rd save.(2nd save same as 1st)
    '3.tiff' was successfully uploaded
    01-Sep-2006 08:37 AM Erik Meitner:
    Relaunching the Finder will restore the icons to the proper state most of the time. One can also use this application to force a refresh of a folders contents: http://www.brockerhoff.net/nudge/index.html
    I am not sure of the internal workings of "Nudge", but it does seem to be the best hack available to get the Finder to work properly.
    01-Sep-2006 08:48 AM Erik Meitner:
    I should also add that this problem is repeatable on 15 different Macs of different models, all running 10.3.9 with the latest updates.
    11-Sep-2006 12:41 PM Erik Meitner:
    Problem is repeatable on all 6 of our Machines running Tiger(10.4.7) with all the latest updates.
    20-Sep-2006 02:25 PM Erik Meitner:
    The "Nudge" tool mentioned previously does not always work especially for some users who experience this problem the worst. It will only restore the file to a "ghosted" icon that is still inaccessable.
    Moving the file to a different folder and deleting the resource fork file is the only solution in this case.

    OS X 10.4.8 update has not altered the behavior of this bug. The Office 2004 11.30 update seems to have caused the problem to happen less frequently. I would say it now happens half as frequently.

  • Expdp fails to create .dmp files in NFS mount point in solaris 10,Oracle10g

    Dear folks,
    I am facing a wierd issue while doing expdp with NFS mount point. Kindly help me on this.
    ===============
    expdp system/manager directory=exp_dumps dumpfile=u2dw.dmp schemas=u2dw
    Export: Release 10.2.0.4.0 - 64bit Production on Wednesday, 31 October, 2012 17:06:04
    Copyright (c) 2003, 2007, Oracle. All rights reserved.
    Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 - 64bit Production
    With the Partitioning, OLAP, Data Mining and Real Application Testing options
    ORA-39001: invalid argument value
    ORA-39000: bad dump file specification
    ORA-31641: unable to create dump file "/backup_db/dumps/u2dw.dmp"
    ORA-27040: file create error, unable to create file
    SVR4 Error: 122: Operation not supported on transport endpoint
    I have mounted like this:
    mount -o hard,rw,noac,rsize=32768,wsize=32768,suid,proto=tcp,vers=3 -F nfs 172.20.2.204:/exthdd /backup_db
    NFS=172.20.2.204:/exthdd
    given read,write grants to public as well as specific user

    782011 wrote:
    Hi sb92075,
    Thanks for ur reply. pls find the below. I am able to touch the files while exporting log files also creating having the error msg as i showed in previous post.
    # su - oracle
    Sun Microsystems Inc. SunOS 5.10 Generic January 2005
    You have new mail.
    oracle 201> touch /backup_db/dumps/u2dw.dmp.test
    oracle 202>I contend that Oracle is too dumb to lie & does not mis-report reality
    27040, 00000, "file create error, unable to create file"
    // *Cause:  create system call returned an error, unable to create file
    // *Action: verify filename, and permissions                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                               

  • NFS Adapter not picking up file

    Hi Guru's,
    Scenario is File(NFS)  to File(FTP) without using repository part.
    The NFS directory name has been maintained correctly. File name is also correct. Channel is active. Checked in RWB also.
    All configuration settings are fine. But NFS adapter is not picking up the file and also not showing any error.
    Tried following thing but not working
    1) Manually deactivating and activating channel in Directory. 
    2) Stop start channel in Communication channel monitoring.
    3) Put exact file name in Communication channel.
    4) Testing mode changed from Delete to Test and again test to delete.
    5) File type binary to text.

    Hello,
    Verify the below mentioned points:
    a. Is the file adapter locked? ....Visual Admin (till PI7.0) or NWA(for PI7.1 and above) . Entry comes as $XIAdapter*
    b. Check permission / authorization to read/ delete the file?
    c. Adapter Services are active
    -Rahul

  • Solaris NFS not able to retrieve a specific version in a folder

    I have a folder with a VersionDescription associated with it. When I open the file through NFS on Solaris or run a "ls" command, it gave this error and never responded:
    NFS server hostname not responding still trying
    If the folder contains the Family, there is no problem. I verified that the VersionDescription association with the folder is working in Windows. Does any know if NFS support VersionDescription?

    I have a folder with a VersionDescription associated with it. When I open the file through NFS on Solaris or run a "ls" command, it gave this error and never responded:
    NFS server hostname not responding still trying
    If the folder contains the Family, there is no problem. I verified that the VersionDescription association with the folder is working in Windows. Does any know if NFS support VersionDescription?

  • NFS: server reboot - client problem with "file handle" - no unmont

    Hello,
    when I restart a NFS-server and a client has a NFS-share on it still mounted, that NFS-share becomes unusable on the client: everything causes the client complain about an old "file handle", even trying to unmount that share!
    So my question is: How to deal gracefully with a NFS-server reboot and avoiding this file handle problem?
    Apart from going to each client and unmount the NFS-share by hand?
    Thanks!

    Do I need to Give these values manually again as they are moved from Env to other?
    When we transport ID objects, the values stored in the channel get lost and have to be manually enetered......it happens even for DEV --> QA environment.
    Now your channel should reflect PROD-related data.
    2- Why my Receiver communiation channel coimg as
    Communication Channel | PRD_400 | RFC_QA3_Receiver:
    The channel name does not depend on the environment it is present....verify the name once again!
    Regards,
    Abhishek.

  • [SOLVED]Can´t mount nfs shares!

    Hello,
    I am trying to get up a fileserver. I installed an old PC with Arch and configerd it over SSH. I followed the wiki over NFS and NFSv4. But everytime I want to mount a share that is in the physical map /home/jozef/shares/downloads, I get the following error:
    mount.nfs4: timeout set for Mon Aug 24 14:36:57 2009
    mount.nfs4: text-based options: 'clientaddr=192.168.0.100,addr=192.168.0.111'
    mount.nfs4: mount(2): Permission denied mount.nfs4: access denied by server while mounting 192.168.0.111:/shares/downloads
    I tried to mount it with the following command:
    mount -v -t nfs4 192.168.0.111:/shares/downloads /media/Downloads-Server
    Lets call the fileserver fileserver and my workstation Gamepc. Here I will give you some files that I think are needed, from fileserver and Gamepc are there other files you need, I will post them.
    /etc/exports on fileserver
    # /etc/exports
    # List of directories exported to NFS clients. See exports(5).
    # Use exportfs -arv to reread.
    # Example for NFSv2 and NFSv3:
    # /srv/home hostname1(rw,sync) hostname2(ro,sync)
    # Example for NFSv4:
    # /srv/nfs4 hostname1(rw,sync,fsid=0)
    # /srv/nfs4/home hostname1(rw,sync,nohide)
    # Using Kerberos and integrity checking:
    # /srv/nfs4 gss/krb5i(rw,sync,fsid=0,crossmnt)
    # /srv/nfs4/home gss/krb5i(rw,sync,nohide)
    /shares 192.168.0.12(rw,fsid=0,no_subtree_check,async,no_root_squash)
    /shares/downloads 192.168.0.12(rw,no_subtree_check,async,no_root_squash,nohide)
    /etc/hosts.allow on fileserver
    # /etc/hosts.allow
    sshd: 192.168.0.0/255.255.255.0
    #nfsd: 192.168.0.0/255.255.255.255
    #portmap: 192.168.0.0/255.255.255.255
    #mountd: 192.168.0.0/255.255.255.255
    nfsd: ALL
    portmap: ALL
    mountd: ALL
    # End of file
    /etc/hosts.deny on fileserver
    # /etc/hosts.deny
    ALL: ALL: DENY
    # End of file
    /etc/conf.d/nfs-common on fileserver
    STATD_OPTS="--no-notify"
    /etc/conf.d/nfs-server on fileserver
    STATD_OPTS="--no-notify"
    /etc/rc.conf on fileserver
    # /etc/rc.conf - Main Configuration for Arch Linux
    # LOCALIZATION
    # LOCALE: available languages can be listed with the 'locale -a' command
    # HARDWARECLOCK: set to "UTC" or "localtime"
    # USEDIRECTISA: use direct I/O requests instead of /dev/rtc for hwclock
    # TIMEZONE: timezones are found in /usr/share/zoneinfo
    # KEYMAP: keymaps are found in /usr/share/kbd/keymaps
    # CONSOLEFONT: found in /usr/share/kbd/consolefonts (only needed for non-US)
    # CONSOLEMAP: found in /usr/share/kbd/consoletrans
    # USECOLOR: use ANSI color sequences in startup messages
    LOCALE="en_US.utf8"
    HARDWARECLOCK=""
    USEDIRECTISA="no"
    TIMEZONE=""
    KEYMAP="us"
    CONSOLEFONT=
    CONSOLEMAP=
    USECOLOR="yes"
    # HARDWARE
    # MOD_AUTOLOAD: Allow autoloading of modules at boot and when needed
    # MOD_BLACKLIST: Prevent udev from loading these modules
    # MODULES: Modules to load at boot-up. Prefix with a ! to blacklist.
    # NOTE: Use of 'MOD_BLACKLIST' is deprecated. Please use ! in the MODULES array.
    MOD_AUTOLOAD="yes"
    #MOD_BLACKLIST=() #deprecated
    MODULES=()
    # Scan for LVM volume groups at startup, required if you use LVM
    USELVM="no"
    # NETWORKING
    # HOSTNAME: Hostname of machine. Should also be put in /etc/hosts
    HOSTNAME="myhost"
    # Use 'ifconfig -a' or 'ls /sys/class/net/' to see all available interfaces.
    # Interfaces to start at boot-up (in this order)
    # Declare each interface then list in INTERFACES
    # - prefix an entry in INTERFACES with a ! to disable it
    # - no hyphens in your interface names - Bash doesn't like it
    # DHCP: Set your interface to "dhcp" (eth0="dhcp")
    # Wireless: See network profiles below
    #Static IP example
    #eth0="dhcp"
    eth0="dhcp"
    INTERFACES=(eth0)
    # Routes to start at boot-up (in this order)
    # Declare each route then list in ROUTES
    # - prefix an entry in ROUTES with a ! to disable it
    gateway="default gw 192.168.0.1"
    ROUTES=(!gateway)
    # Enable these network profiles at boot-up. These are only useful
    # if you happen to need multiple network configurations (ie, laptop users)
    # - set to 'menu' to present a menu during boot-up (dialog package required)
    # - prefix an entry with a ! to disable it
    # Network profiles are found in /etc/network.d
    # This now requires the netcfg package
    #NETWORKS=(main)
    # DAEMONS
    # Daemons to start at boot-up (in this order)
    # - prefix a daemon with a ! to disable it
    # - prefix a daemon with a @ to start it up in the background
    DAEMONS=(syslog-ng !network rpcbind nfs-common nfs-server netfs crond !xinetd !samba sshd transmissiond)
    /etc/rc.conf on Gamepc
    # /etc/rc.conf - Main Configuration for Arch Linux
    # LOCALIZATION
    # LOCALE: available languages can be listed with the 'locale -a' command
    # HARDWARECLOCK: set to "UTC" or "localtime"
    # USEDIRECTISA: use direct I/O requests instead of /dev/rtc for hwclock
    # TIMEZONE: timezones are found in /usr/share/zoneinfo
    # KEYMAP: keymaps are found in /usr/share/kbd/keymaps
    # CONSOLEFONT: found in /usr/share/kbd/consolefonts (only needed for non-US)
    # CONSOLEMAP: found in /usr/share/kbd/consoletrans
    # USECOLOR: use ANSI color sequences in startup messages
    LOCALE="en_US.utf8"
    HARDWARECLOCK="localtime"
    USEDIRECTISA="no"
    TIMEZONE="Europe/Amsterdam"
    KEYMAP="us"
    CONSOLEFONT=
    CONSOLEMAP=
    USECOLOR="yes"
    # HARDWARE
    # MOD_AUTOLOAD: Allow autoloading of modules at boot and when needed
    # MOD_BLACKLIST: Prevent udev from loading these modules
    # MODULES: Modules to load at boot-up. Prefix with a ! to blacklist.
    # NOTE: Use of 'MOD_BLACKLIST' is deprecated. Please use ! in the MODULES array.
    MOD_AUTOLOAD="yes"
    #MOD_BLACKLIST=() #deprecated
    MODULES=()
    # Scan for LVM volume groups at startup, required if you use LVM
    USELVM="no"
    # NETWORKING
    # HOSTNAME: Hostname of machine. Should also be put in /etc/hosts
    HOSTNAME="GAMEPC"
    # Use 'ifconfig -a' or 'ls /sys/class/net/' to see all available interfaces.
    # Interfaces to start at boot-up (in this order)
    # Declare each interface then list in INTERFACES
    # - prefix an entry in INTERFACES with a ! to disable it
    # - no hyphens in your interface names - Bash doesn't like it
    # DHCP: Set your interface to "dhcp" (eth0="dhcp")
    # Wireless: See network profiles below
    #Static IP example
    #eth0="dhcp"
    eth0="dhcp"
    INTERFACES=(eth0)
    # Routes to start at boot-up (in this order)
    # Declare each route then list in ROUTES
    # - prefix an entry in ROUTES with a ! to disable it
    gateway="default gw 192.168.0.1"
    ROUTES=(!gateway)
    # Enable these network profiles at boot-up. These are only useful
    # if you happen to need multiple network configurations (ie, laptop users)
    # - set to 'menu' to present a menu during boot-up (dialog package required)
    # - prefix an entry with a ! to disable it
    # Network profiles are found in /etc/network.d
    # This now requires the netcfg package
    #NETWORKS=(main)
    # DAEMONS
    # Daemons to start at boot-up (in this order)
    # - prefix a daemon with a ! to disable it
    # - prefix a daemon with a @ to start it up in the background
    DAEMONS=(syslog-ng !network rpcbind nfs-common @netfs crond @hal @fam @samba @alsa @cups)
    /etc/hosts.allow on Gamepc
    # /etc/hosts.allow
    sshd:ALL
    # End of file
    /etc/hosts.deny on Gamepc
    # /etc/hosts.deny
    ALL: ALL: DENY
    # End of file
    If someone could help me, I would be pleased.
    With kind regards,
    Jozef00
    Last edited by jozef00 (2009-09-01 15:28:07)

    Are you sure that the nfs module is loaded? Do
    lsmod | grep nfs
    to check. If "nfs" doesn't appear then that's your problem. Do:
    sudo modprobe nfs
    and continue.
    EDIT: Also, I'll assume both your server and your client machine are connected to the network. You have the network DAEMON "!"-disabled in each machine's /etc/rc.conf. So I'm assuming you're connecting them to the network in some other way, and have verified that you can, for example, ssh from one machine to the other.
    Third, you're sure that /share exists on your server machine, that a (presumably empty) directory /share/downloads exists inside it, that another folder is mounted on top of that empty directory? When you type "mount" on the server machine there should be a line containing "on /share/downloads".
    Fourth, if your client's ip address is 192.168.0.100 (this will be displayed when you do "ifconfig") then the /etc/exports on your server has been changed to read:
    /shares 192.168.0.100(rw,fsid=0,no_subtree_check,async,no_root_squash)
    /shares/downloads 192.168.0.100(rw,no_subtree_check,async,no_root_squash,nohide)
    Then you either typed "exportfs -rf" on the server, or stopped and restarted nfs-server. (I'd stop nfs-server, nfs-common, and rpcbind, then start them up again in reverse order, to be sure.)
    STATD_OPTS has no effect in /etc/conf.d/nfs-server. Also, you don't need to use statd if you're using nfs4. So you could change /etc/conf.d/nfs-common to read:
    NEED_STATD=no
    NEED_IDMAPD=yes
    and /etc/conf.d/server and can be blank.
    (Then restart everything, as described above.)
    If you're using the latest arch packages then portmap has been replaced by rpcbind (which you do have in your DAEMONS line on the server). So you need to change the "portmap:" line in your server's /etc/hosts.allow to "rpcbind:". I also have a "lockd:" line in my /etc/hosts.allow, in addition to what you have. It's probably not the source of your present troubles, and I'm not sure it's necessary. But you could try adding it.
    Finally, did you add/change a "Domain = ..." line in the /etc/idmapd.conf on your server, and a corrsponding line in the /etc/idmapd.conf on your client?
    Last edited by Profjim (2009-08-25 12:51:48)

Maybe you are looking for