NFS shares and df problem

I have two nfs shares mounted on bootup, "10.0.0.118:/warehouse" and "10.0.0.118:/lataukset". However when I use df, it always displays something like this:
udev 10M 0 10M 0% /dev
run 10M 260K 9,8M 3% /run
/dev/sda1 30G 5,7G 23G 21% /
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/warehouse 5,4T 3,3T 1,9T 65% /mnt/warehouse
10.0.0.118:/lataukset 908G 84G 778G 10% /mnt/lataukset
So same entry is shown multiple times.
/etc/fstab
10.0.0.118:/lataukset /mnt/lataukset nfs4 defaults,user,noauto,comment=systemd.automount 0 0
10.0.0.118:/warehouse /mnt/warehouse nfs4 defaults,user,noauto,comment=systemd.automount 0 0
What could be the reason for this and is there a fix for it?
Thanks.

My best guess is "comment=systemd.automount" and do u have any reaon to use "noauto" ?
i always mount nfs4 shares with "defaults,async,user,proto=tcp,soft,intr" ( i m not realy sure what is included and what is not in "defaults so i just put those there, they dont byte so far Oo) u can only think about async option ( no problems so far ~1 year but who knows i think it was not recommended at least for nfs v3)
well that is it good luck

Similar Messages

  • Share and save problem after latest update

    hello,
    i have share and save problem after latest update - i have installed ps touch on nexus 5 about 2 months ago and until yesterday everything was good but after latest update i can't save and share images in ps touch ( i do reinstall this app  but problem is still  and app doesn't work properly   ) how  can i downgrade to the old version ? Could you fix this problem quickly ?

    i use app which called "aLogcat"
    --------- beginning of /dev/log/main
    I/dalvikvm(24476): Enabling JNI app bug workarounds for target SDK version 11...
    D/dalvikvm(24476): GC_CONCURRENT freed 174K, 2% free 17054K/17264K, paused 2ms+2ms, total 15ms
    W/InputEventReceiver(24476): Attempted to finish an input event but the input event receiver has already been disposed.
    W/InputEventReceiver(24476): Attempted to finish an input event but the input event receiver has already been disposed.
    W/InputEventReceiver(24476): Attempted to finish an input event but the input event receiver has already been disposed.
    --------- beginning of /dev/log/system
    W/ViewRootImpl(24476): Dropping event due to root view being removed: MotionEvent { action=ACTION_MOVE, id[0]=0, x[0]=38.0, y[0]=-115.0, toolType[0]=TOOL_TYPE_FINGER, buttonState=0, metaState=0, flags=0x0, edgeFlags=0x0, pointerCount=1, historySize=2, eventTime=27743080, downTime=27743066, deviceId=4, source=0x1002 }
    W/ViewRootImpl(24476): Dropping event due to root view being removed: MotionEvent { action=ACTION_UP, id[0]=0, x[0]=38.0, y[0]=-115.0, toolType[0]=TOOL_TYPE_FINGER, buttonState=0, metaState=0, flags=0x0, edgeFlags=0x0, pointerCount=1, historySize=0, eventTime=27743086, downTime=27743066, deviceId=4, source=0x1002 }
    D/dalvikvm(24476): GC_FOR_ALLOC freed 122K, 1% free 17271K/17428K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 10K, 1% free 17434K/17604K, paused 7ms, total 7ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 65K, 2% free 17690K/17928K, paused 8ms, total 8ms
    I/dalvikvm-heap(24476): Grow heap (frag case) to 17.615MB for 328336-byte allocation
    D/dalvikvm(24476): GC_FOR_ALLOC freed 0K, 2% free 18010K/18252K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 18011K/18252K, paused 9ms, total 9ms
    I/dalvikvm-heap(24476): Grow heap (frag case) to 18.073MB for 479536-byte allocation
    D/dalvikvm(24476): GC_FOR_ALLOC freed 0K, 2% free 18479K/18724K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 2% free 18481K/18724K, paused 2ms+1ms, total 10ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 6ms
    I/dalvikvm-heap(24476): Grow heap (frag case) to 18.532MB for 479536-byte allocation
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 18949K/19196K, paused 9ms, total 9ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 2% free 19418K/19668K, paused 1ms+1ms, total 10ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 7ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 19592K/19844K, paused 8ms, total 8ms
    I/dalvikvm-heap(24476): Grow heap (frag case) to 20.049MB for 933136-byte allocation
    D/dalvikvm(24476): GC_FOR_ALLOC freed 0K, 2% free 20503K/20756K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 2% free 21416K/21668K, paused 1ms+1ms, total 12ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 50K, 2% free 22351K/22580K, paused 7ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 5K, 1% free 23900K/24140K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 1% free 25824K/26084K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 3411K, 14% free 24528K/28204K, paused 12ms, total 13ms
    D/dalvikvm(24476): GC_CONCURRENT freed 2814K, 14% free 24256K/28204K, paused 2ms+1ms, total 11ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 4ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 9266K, 1% free 17270K/17424K, paused 10ms, total 10ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 15K, 1% free 17575K/17748K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 22K, 2% free 17875K/18072K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 1K, 2% free 18195K/18396K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 18517K/18720K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 18838K/19044K, paused 11ms, total 11ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 19159K/19368K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 19480K/19692K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 19974K/20192K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 20616K/20840K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 21431K/21664K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 2% free 22714K/22960K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 182K, 2% free 24004K/24432K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 8478K, 33% free 17593K/26228K, paused 14ms, total 14ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 18K, 32% free 17895K/26228K, paused 12ms, total 12ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 18K, 31% free 18198K/26228K, paused 11ms, total 11ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 1K, 30% free 18519K/26228K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 29% free 18840K/26228K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 26% free 19482K/26228K, paused 2ms+1ms, total 14ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 10ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 23% free 20297K/26228K, paused 2ms+1ms, total 15ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 11ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 19% free 21260K/26228K, paused 10ms, total 10ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 15% free 22543K/26228K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 182K, 9% free 23979K/26228K, paused 21ms, total 22ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 8485K, 33% free 17598K/26132K, paused 13ms, total 14ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 25K, 32% free 17894K/26132K, paused 12ms, total 12ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 12K, 31% free 18204K/26132K, paused 12ms, total 12ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 1K, 30% free 18524K/26132K, paused 12ms, total 12ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 28% free 18845K/26132K, paused 8ms, total 9ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 26% free 19487K/26132K, paused 1ms+0ms, total 9ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 7ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 23% free 20129K/26132K, paused 9ms, total 9ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 20% free 21091K/26132K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 15% free 22375K/26132K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 9K, 9% free 24011K/26132K, paused 14ms, total 14ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 2147K, 10% free 23855K/26456K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_CONCURRENT freed 8776K, 35% free 17337K/26304K, paused 3ms+2ms, total 20ms
    W/InputEventReceiver(24476): Attempted to finish an input event but the input event receiver has already been disposed.
    W/InputEventReceiver(24476): Attempted to finish an input event but the input event receiver has already been disposed.
    W/ViewRootImpl(24476): Dropping event due to root view being removed: MotionEvent { action=ACTION_UP, id[0]=0, x[0]=564.0, y[0]=-50.0, toolType[0]=TOOL_TYPE_FINGER, buttonState=0, metaState=0, flags=0x0, edgeFlags=0x0, pointerCount=1, historySize=0, eventTime=28190972, downTime=28190972, deviceId=4, source=0x1002 }
    D/dalvikvm(24476): GC_CONCURRENT freed 186K, 31% free 17626K/25476K, paused 2ms+1ms, total 12ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 4ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 63K, 30% free 17884K/25476K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 3K, 29% free 18201K/25476K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 27% free 18696K/25476K, paused 2ms+1ms, total 12ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 9ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 25% free 19190K/25476K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 23% free 19831K/25476K, paused 9ms, total 9ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 19% free 20794K/25476K, paused 2ms+0ms, total 12ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 8ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 14% free 22078K/25476K, paused 1ms+2ms, total 14ms
    D/dalvikvm(24476): WAIT_FOR_CONCURRENT_GC blocked 12ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 7% free 23855K/25476K, paused 10ms, total 10ms
    D/dalvikvm(24476): GC_CONCURRENT freed <1K, 6% free 24176K/25476K, paused 2ms+1ms, total 14ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 6440K, 26% free 19562K/26216K, paused 9ms, total 10ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 93K, 23% free 20207K/26216K, paused 10ms, total 10ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed 11K, 21% free 20935K/26216K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 18% free 21673K/26216K, paused 9ms, total 9ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 15% free 22499K/26216K, paused 8ms, total 8ms
    D/dalvikvm(24476): GC_FOR_ALLOC freed <1K, 10% free 23695K/26216K, paused 9ms, total 9ms

  • Pcnfsd authetication in NFS share

    I have a G5 10.4.8 serving a NFS share. The one client is a Win95 box running NFS Maestro . It has worked before but in rebuilding it, I'm stuck. The client requires pcnfsd authentication but inetd.conf states that (rpc.) pcnfsd is not yet implemented in OS X. NFS Manager has nothing as well. Is there a way to fix this?
    G5   Mac OS X (10.4.8)   Not OSX Server

    This is frustrating!.. I've managed to get that working but i have a different issue now ..here is what i have done in details:
    Server: MAC OSX 10.6.8.
    Server was standalone and then I bounded to the AD then promoted to Open Directory Master without kerberos realm - as it is the AD is the one that holds the accounts.. that's how it should be correct ?
    Disk utility: I mounted the NFS share and in WGM i enabled file sharing on the NFS share via AFP, and now it shows under WGM - Home tab as: afp://xserve.mydomain.com/homes
    - for the clients, I bounded them to OD first for MCX and then to AD.
    Directory Utility: settings for both "client & server"
    *Create mobile account at login - false
    *Force local home directory on startup disk - false
    *Une UNC path - True with AFP protocol.
    Server Admin:
    * AFP is enabled on the NFS share "homes" and its auto-mounted.
    * Open Directory Master:
    - LDAP Server is running
    - Password Server is running.
    - Kerberos is stopped.
    Workgroup Manager:
    * I selected the test user "adtest" and assigned the home folder which is:
    Home URL: afp://xserve.domain.com/homes/adtest
    Full Path: /Network/Servers/xserve.domain.com/homes/adtest
    and when i click on Create Home Now, it did create the user home directory under the NFS share which is auto-mounted
    Active Directory Server:
    under the adtest user - profile tab, i see: \\xserve.domain.com\homes\adtest
    Problem:
    - when i try to login with adtest user from the client, i get the error message:
    "You are unable to login in o the user account "adtest" at this time - logging in to the account failed because an error occurred."
    Troubleshooting:
    1- logged in with local admin account and typed id adtest in terminal.. it shows all user attributes and groups, which means the machine is bound correctly to both AD & OD
    2- when I change the home for the user to the the default "/Users".. i can login just fine with the adtest account.. does that look like its a permissions issue ?
    Thanks again for your help.

  • Have a MacMini that is dropping NFS Share connections, Using cifs://servername/path however the NFS Share is dropped everyday without fail.

    Have a MacMini that is dropping NFS Share connections, Using cifs://servername/path however the NFS Share is dropped everyday without fail.  Is there a better way to setup this share?  I've got 5 other Macs that use the same NFS Share and none of them are dropping the shared connection.  This is on OSX 10.9.x Maverick.

    cifs:// is not NFS cifs:// is effectively SMB i.e. Windows filesharing, NFS is Unix filesharing, furthermore cifs:// is a way of forcing Mavericks to use SMB version 1 protocol. Normally you would use smb:// which will default to trying to use SMB version 2 protocol.
    Saying all that, SMB version 1, SMB version 2, and NFS should all normally work fine. It is possible that the server may be configured to disconnect idle users although one would expect this to happen with other users as well. It is perhaps more likely that your problem Mavericks Mac maybe going to sleep periodically, while asleep the connection to the server maybe being lost. Apple's own server software has specific cleverness to allow clients to sleep and resume their connection when they wake.
    Therefore it might be worth checking your Energy Saver settings on this Mac and disabling computer sleep.

  • Use NAS (NFS) share as storage? Can't make shares.

    Hi,
    Trying to make shares on a network storage, nas, connected via nfs but i noticed that its greyed out. Is it supposed to only use disks or iSCSI för shares?

    This is frustrating!.. I've managed to get that working but i have a different issue now ..here is what i have done in details:
    Server: MAC OSX 10.6.8.
    Server was standalone and then I bounded to the AD then promoted to Open Directory Master without kerberos realm - as it is the AD is the one that holds the accounts.. that's how it should be correct ?
    Disk utility: I mounted the NFS share and in WGM i enabled file sharing on the NFS share via AFP, and now it shows under WGM - Home tab as: afp://xserve.mydomain.com/homes
    - for the clients, I bounded them to OD first for MCX and then to AD.
    Directory Utility: settings for both "client & server"
    *Create mobile account at login - false
    *Force local home directory on startup disk - false
    *Une UNC path - True with AFP protocol.
    Server Admin:
    * AFP is enabled on the NFS share "homes" and its auto-mounted.
    * Open Directory Master:
    - LDAP Server is running
    - Password Server is running.
    - Kerberos is stopped.
    Workgroup Manager:
    * I selected the test user "adtest" and assigned the home folder which is:
    Home URL: afp://xserve.domain.com/homes/adtest
    Full Path: /Network/Servers/xserve.domain.com/homes/adtest
    and when i click on Create Home Now, it did create the user home directory under the NFS share which is auto-mounted
    Active Directory Server:
    under the adtest user - profile tab, i see: \\xserve.domain.com\homes\adtest
    Problem:
    - when i try to login with adtest user from the client, i get the error message:
    "You are unable to login in o the user account "adtest" at this time - logging in to the account failed because an error occurred."
    Troubleshooting:
    1- logged in with local admin account and typed id adtest in terminal.. it shows all user attributes and groups, which means the machine is bound correctly to both AD & OD
    2- when I change the home for the user to the the default "/Users".. i can login just fine with the adtest account.. does that look like its a permissions issue ?
    Thanks again for your help.

  • Mounting an NFS share

    Hi,
    I'm attempting to mount an NFS share and having no success.  Regardless of the settings I try, the Finder still denies me access to the NFS share, even though it mounts fine.  I seem to have no read or write access to the share.
    I've tried exporting the share (in /etc/exports on the server machine) in two ways: with
    /home/REDACTED/share     REDACTED/28(rw,sync,all_squash)
    And
    /home/REDACTED/share     REDACTED/28(rw,sync,insecure,all_squash,anonuid=1001,anongid=1001)
    In the second example, the anonuid and anongid are those of the shared folder's owner and group.  I added "insecure" because a how-to on the web claims that OS X won't work with any shares that don't have this specified.
    With either of these settings applied, Disk Utility verifies the existence of the share, and mounts it.  However, I can neither read files within, or add files to, the shared folder.  The error produced is:
    The folder “share” can’t be opened because you don’t have permission to see its contents.
    I have tried the following Advanced Mount Parameters, each to no effect:
    nodev resvport nolocks locallocks intr soft wsize=32768 rsize=3276
    nodev nosuid resvport nolocks locallocks intr soft wsize=32768 rsize=3276 ro
    nodev,nosuid,resvport,nolocks,locallocks,intr,soft,wsize=32768,rsize=3276
    nodev,nosuid,resvport,nolocks,locallocks,intr,soft,wsize=32768,rsize=3276 ro
    resvport,nolocks,locallocks,intr,soft,wsize=32768,rsize=3276
    -i,-s,-w=32768,-r32768
    -P
    I'd rather not employ SAMBA, and the Apple File Sharing package for my server's OS (Ubuntu 11.10) appears to be bugged currently.  Besides, NFS would be a far neater solution.
    Any helpful advice?
    S.

    Scotch_Brawth wrote:
    I've simply come at NFS as being the most appropriate file-sharing implementation for my needs - it supports automatic mounting at boot using tech native to both my Linux OS and OS X. 
    That is part of the problem. NFS is designed for environments where all servers are mounted (by root) at boot time and permissions are managed via NIS or LDAP. That is the default setting. If you are using something else, it requires some hacking.
    I've had SAMBA working in the past, but I guess a certain air of contamination creeps in when using a Windows protocol to allow interaction between two UNIXy systems.
    Plus, you would now have two different 3rd party reverse-engineered reimplementations of a foreign protocol.
    AFP would be great, but despite receiving support on the Ubuntu forums and IRC, I failed to get it to work - it may be bugged; which would not be surprising, as 11.10 (with Kernel 3) has proved problematic in several other ways.
    Perhaps Ubuntu is targeted more towards desktop rather than server usage. About the time I last played with NFS, I also played with Netatalk - with disastrous results. Supposedly Netatalk is better now. It's authors would be more than happy to sell you a support package.
    I did use the default settings - they failed to allow a working NFS share.  I then applied the variety of settings as recommended by apparently knowledgeable people.  Still no success.  I have read that UID/GID settings are an important aspect of NFS, but the issue in this case (as far as I understand it) is that all UID/GIDs below 1000 are privileged in Ubuntu 11.10, whilst on OS X these are below 501.  So, the choice is either to give the shared folder owner a privileged UID/GID pair, or change the UID/GID of my Mac users to meet the NFS servers needs - not something I'm happy to do for so small a gain.
    You can create a throwaway account on the Mac and just reset the GID/UID to values equal to an account on the Linux machine. That would establish that it is properly working in the default configuration. Then you could edit /etc/idmapd.conf.
    For that reason, I use the "all_squash" option, because the share in question is not for anything remotely critical and the data to be transferred and stored is both worthless and transitory.
    Since all_squash maps everything to nobody, you would have to hack up the permissions on the server to make everything world writeable. I think it will work with /etc/idmapd.conf and without all_squash.
    I know nothing about NFS other than that its capabilities and integration meet my needs.
    Just what are your needs? If the data is worthless and not critical then Netatalk might be the best option. If you can't get that to work, you could try MacFUSE on the Mac side and mount over sshfs. That is normally what I do. It isn't all that reliable, but you don't seem to require that.
    What information I did find regarding OS X and NFS was that there were peculiarities that required certain settings to be present on the server and the client respectively - for example, OS X apparently requires "insecure" to be set as an option, or it simply won't connect properly.  I don't know why, but I have no choice to trust to the advice of others in this case, until I have sufficient grasp to take care of the whole thing myself.
    This goes back to the expectation that NFS expects to be always connected and mounted by root. Apple sells very few desktop machines anymore so it assumes a different, user-centered environment. You could use "insecure" on the server side to allow connections from "insecure" ports > 1024 that a regular users can connect with via the Finder. You could use the terminal with "sudo mount_nfs -o resvport" to tell the Mac to use the root user to connect via a secure port instead.
    If you genuinely think you're able to help, then I'm happy to hear your advice.  What would you recommend?
    I appreciate your meeting me halfway. I think all you really need is /etc/idmapd.conf without all_squash. Then you could setup AutoFS and you could use NFS in a modern environment without even bothering to mount it.

  • New files and folders on a Linux client mounting a Windows 2012 Server for NFS share do not inherit Owner and Group when SetGID bit set

    Problem statement
    When I mount a Windows NFS service file share using UUUA and set the Owner and Group, and set the SetGID bit on the parent folder in a hierarchy. New Files and folders inside and underneath the parent folder do not inherit the Owner and Group permissions
    of the parent.
    I am given to understand from this Microsoft KnowledgeBase article (http://support.microsoft.com/kb/951716/en-gb) the problem is due to the Windows implmentation of NFS Services not supporting the Solaris SystemV or BSD grpid "Semantics"
    However the article says the same functionality can acheived by using ACE Inheritance in conjunction with changing the Registry setting for "KeepInheritance" to enable Inheritance propagation of the Permissions by the Windows NFS Services.
    1. The Precise location of the "KeepInheritance" DWORD key appears to have "moved" in  Windows Server 2012 from a Services path to a Software path, is this documented somewhere? And after enabling it, (or creating it in the previous
    location) the feature seems non-functional. Is there a method to file a Bug with Microsoft for this Feature?
    2. All of the references on demonstrating how to set an ACE to achieve the same result "currently" either lead to broken links on Microsoft technical websites, or are not explicit they are vague or circumreferential. There are no plain Examples.
    Can an Example be provided?
    3. Is UUUA compatible with the method of setting ACE to acheive this result, or must the Linux client mount be "Mapped" using an Authentication source. And could that be with the new Flat File passwd and group files in c:\windows\system32\drivers\etc
    and is there an Example available.
    Scenario:
    Windows Server 2012 Standard
    File Server (Role)
    +- Server for NFS (Role) << -- installed
    General --
    Folder path: F:\Shares\raid-6-array
    Remote path: fs4:/raid-6-array
    Protocol: NFS
    Authentication --
    No server authentication
    +- No server authentication (AUTH_SYS)
    ++- Enable unmapped user access
    +++- Allow unmapped user access by UID/GID
    Share Permissions --
    Name: linux_nfs_client.host.edu
    Permissions: Read/Write
    Root Access: Allowed
    Encoding: ANSI
    NTFS Permissions --
    Type: Allow
    Principal: BUILTIN\Administrators
    Access: Full Control
    Applies to: This folder only
    Type: Allow
    Principal: NT AUTHORITY\SYSTEM
    Access: Full Control
    Applies to: This folder only
    -- John Willis, Facebook: John-Willis, Skype: john.willis7416

    I'm making some "major" progress on this problem.
    1. Apparently the "semantics" issue to honor SGID or grpid in NFS on the server side or the client side has been debated for some time. It also existed as of 2009 between Solaris nfs server and Linux nfs clients. The Linux community defaulted to declaring
    it a "Server" side issue to avoid "Race" conditions between simultaneous access users and the local file system daemons. The client would have to "check" for the SGID and reformulate its CREATE request to specify the Secondary group it would have to "notice"
    by which time it could have changed on the server. SUN declined to fix it.. even though there were reports it did not behave the same between nfs3 vs nfs4 daemons.. which might be because nfs4 servers have local ACL or ACE entries to process.. and a new local/nfs
    "inheritance" scheme to worry about honoring.. that could place it in conflict with remote access.. and push the responsibility "outwards" to the nfs client.. introducing a race condition, necessitating "locking" semantics.
    This article covers that discovery and no resolution - http://thr3ads.net/zfs-discuss/2009/10/569334-CR6894234-improved-sgid-directory-compatibility-with-non-Solaris-NFS-clients
    2. A much Older Microsoft Knowledge Based article had explicit examples of using Windows ACEs and Inheritance to "mitigate" the issue.. basically the nfs client "cannot" update an ACE to make it "Inheritable" [-but-] a Windows side Admin or Windows User
    [-can-] update or promote an existing ACE to "Inheritable"
    Here are the pertinent statements -
    "In Windows Services for UNIX 2.3, you can use the KeepInheritance registry value to set inheritable ACEs and to make sure that these ACEs apply to newly created files and folders on NFS shares."
    "Note About the Permissions That Are Set by NFS Clients
    The KeepInheritance option only applies ACEs that have inheritance enabled. Any permissions that are set by an NFS client will
    only apply to that file or folder, so the resulting ACEs created by an NFS client will
    not have inheritance set."
    "So
    If you want a folder's permissions to be inherited to new subfolders and files, you must set its permissions from the Windows NFS server because the permissions that are set by NFS clients only apply to the folder itself."
    http://support.microsoft.com/default.aspx?scid=kb;en-us;321049
    3. I have set up a Windows 2008r2 NFS server and mounted it with a Redhat Enteprise Linux 5 release 10 x86_64 server [Oct 31, 2013] and so far this does appear to be the case.
    4. In order to mount and then switch user to a non-root user to create subdirectories and files, I had to mount the NFS share (after enabling Anonymous AUTH_SYS mapping) this is not a good thing, but it was because I have been using UUUA - Unmapped Unix
    User Access Mapping, which makes no attempt to "map" a Unix UID/GID set by the NFS client to a Windows User account.
    To verify the Inheritance of additional ACEs on new subdirectories and files created by a non-root Unix user, on the Windows NFS server I used the right click properties, security tab context menu, then Advanced to list all the ACEs and looked at the far
    Column reflecting if it applied to [This folder only, or This folder and Subdirectories, or This folder and subdirectories and files]
    5. All new Subdirectories and files createdby the non-root user had a [Non-Inheritance] ACE created for them.
    6. I turned a [Non-Inheritance] ACE into an [Inheritance] ACE by selecting it then clicking [Edit] and using the Drop down to select [This folder, subdirs and files] then I went back to the NFS client and created more subdirs and files. Then back to the
    Windows NFS server and checked the new subdirs and folders and they did Inherit the Windows NFS server ACE! - However the UID/GID of the subdirs and folders remained unchanged, they did not reflect the new "Effective" ownership or group membership.
    7. I "believe" because I was using UUUA and working "behind" the UID/GID presentation layer for the NFS client, it did not update that presentation layer. It might do that "if" I were using a Mapping mechanism and mapped UID/GID to Windows User SIDs and
    Group SIDs. Windows 2008r2 no longer has a "simple" Mapping server, it does not accept flat text files and requires a Schema extension to Active Directory just to MAP a windows account to a UID/GID.. a lot of overhead. Windows Server 2012 accepts flat text
    files like /etc/passwd and /etc/group to perform this function and is next on my list of things to see if that will update the UID/GID based on the Windows ACE entries. Since the Local ACE take precedence "over" Inherited ACEs there could be a problem. The
    Inheritance appears to be intended [only] to retain Administrative rights over user created subdirs and files by adding an additional ACE at the time of creation.
    8. I did verify from the NFS client side in Linux that "Even though" the UID/GID seem to reflect the local non-root user should not have the ability to traverse or create new files, the "phantom" NFS Server ACEs are in place and do permit the function..
    reconciling the "view" with "reality" appears problematic, unless the User Mapping will update "effective" rights and ownership in the "view"
    -- John Willis, Facebook: John-Willis, Skype: john.willis7416

  • Windows 2012 R2 - Access problems to NFS shares via IPv6

    Hello,
    we setup some NFS shares on HNAS storage box. The connection to this storage is only possible over IPv6 network.
    My server with Windows 2012 R2 has the feature 'Client for NFS' installed and a 'showmount -e servername/or IPv6 address' works. I can see all shares when i try to open the storagebox in Windows Explorer with the command "\\servername" or "\\IPv6
    address", too. 
    My problem is, when I try to mount the share via "Map network drive" or "mount \\servername\sharename *" I get Error 53 - Network path not found.
    The share is configured that every server in the IPv6 subnet could access it and it works on Unix machines, but not on Windows. 
    It does not matter if i set some option parameters for the "mount" command like sec=sys or anon, everytime error 53.
    I don't know, if the share needs special options for Windows. I hope somebody can help.
    Kind regards

    Hi,
    here is the output of your 2 commands:
    PS C:\Users\admin> get-SmbConnection
    PS C:\Users\admin> get-SmbServerConfiguration
    AnnounceServer                  : False
    AsynchronousCredits             : 64
    AutoShareServer                 : True
    AutoShareWorkstation            : True
    CachedOpenLimit                 : 0
    AnnounceComment                 :
    EnableDownlevelTimewarp         : False
    EnableLeasing                   : True
    EnableMultiChannel              : True
    EnableStrictNameChecking        : True
    AutoDisconnectTimeout           : 0
    DurableHandleV2TimeoutInSeconds : 30
    EnableAuthenticateUserSharing   : False
    EnableForcedLogoff              : True
    EnableOplocks                   : True
    EnableSecuritySignature         : True
    ServerHidden                    : True
    IrpStackSize                    : 15
    KeepAliveTime                   : 2
    MaxChannelPerSession            : 32
    MaxMpxCount                     : 50
    MaxSessionPerConnection         : 16384
    MaxThreadsPerQueue              : 20
    MaxWorkItems                    : 1
    NullSessionPipes                : HydraLsPipe
    NullSessionShares               :
    OplockBreakWait                 : 35
    PendingClientTimeoutInSeconds   : 120
    RequireSecuritySignature        : False
    EnableSMB1Protocol              : True
    EnableSMB2Protocol              : False
    Smb2CreditsMax                  : 2048
    Smb2CreditsMin                  : 128
    SmbServerNameHardeningLevel     : 0
    TreatHostAsStableStorage        : False
    ValidateAliasNotCircular        : True
    ValidateShareScope              : True
    ValidateShareScopeNotAliased    : True
    ValidateTargetName              : True
    EncryptData                     : False
    RejectUnencryptedAccess         : True
    I try to access the file share with Windows Explorer. When I enter \\fe08--babe-face-cafe-dead.ipv6-literal.net (just example), I can see all shares on the storage box. When I double-click the share, I want to access, where an access is definitely allowed,
    it takes a while and then the error "\\fe08--babe-face-cafe-dead.ipv6-literal.net\share is not accessable. You might not have permission to use this network resource. ..." appears.  Of this action a recorded the trace.
    Another way I tried is to open command line an type "mount \\fe08--babe-face-cafe-dead.ipv6-literal.net\share *". After this an error 53 occurs immediately without any wait time.

  • [SOLVED] Problem mounting nfs shares

    Hello everybody,
    I keep getting the following error when I try to mount an NFS share;
    mount.nfs4: an incorrect mount option was specified
    The corresponding line in /etc/fstab is this:
    pc-francois.local:/francois /mnt/pc-francois nfs4 rw,port=2049,noauto,users,intr  0 0
    showmount -e pc-francois.local gives this:
    Export list for pc-francois.local:
    /export/archives 192.168.1.0/24
    /export/portage  192.168.1.0/24
    /export/francois 10.8.0.0/24,192.168.1.0/24
    /export          10.8.0.0/24,192.168.1.0/24
    Finally, the nfs related modules are loaded. lsmod | grep nfs gives this:
    nfsv4                 129119  0
    nfsd                  228113  13
    auth_rpcgss            41857  1 nfsd
    nfs                   118782  1 nfsv4
    lockd                  57996  2 nfs,nfsd
    grace                   1706  2 nfsd,lockd
    sunrpc                170119  19 nfs,nfsd,auth_rpcgss,lockd,nfsv4
    It works perfectly fine from other client not running on arch linux. Like indicated in the howto, I have enabled and started the services rpcbind.service,nfs-client.target and remote-fs.target
    Can somebody explains what is happening ?
    Last edited by FrançoisVal (2015-05-02 16:48:14)

    I finally solved my problem by changing the mount options to  this in fstab:
    pc-francois.local:/francois /mnt/pc-francois nfs4 rw,proto=tcp,vers=4.0,noauto,users,intr  0 0
    One week to find it... Maybe it wants to use nfs 4.1 by default.

  • NFS share problem

    When I start up the NFS shares I get an error:
    "SC[SUNW.nfs:3.1,nfs-rg,nfs-res,nfs_probe]: stat of share path /global/qfsnfs1 failed: Value too large for defined data type. Terminated"
    The qfs file system /global/qfsnfs1is 5.4Terabytes and if I turn of the fault monitoring on the NFS resouce I can share it okay.
    Ths is the same system I posted QFS issues with earlier.
    Any ideas??

    When I start up the NFS shares I get an error:
    "SC[SUNW.nfs:3.1,nfs-rg,nfs-res,nfs_probe]: stat of
    share path /global/qfsnfs1 failed: Value too large
    for defined data type. Terminated"
    The qfs file system /global/qfsnfs1is 5.4Terabytes
    and if I turn of the fault monitoring on the NFS
    resouce I can share it okay.
    Ths is the same system I posted QFS issues with
    earlier.
    Any ideas??Did you find a resolution?
    I am experiencing the same thing
    My file system ia 17TB
    when I reduce it to 800GB it works just fine
    Can share manually at 17TB but the SC agent chokes when I make it larger than 1TB
    v440, S9, SC3.1U4,QFS4.5
    Thanks

  • Nautilus NFS shares listed under both Devices and Network

    I have several NFS shares from my home server that I mount using fstab.  I am running a fully up to date installation of Gnome 3.12.  However, as my subject indicates, my NFS shares appear under both Devices and Network in Nautilus.  I've done some searching and not found much about this.  It is redundant and most importantly space wasting to have two lists of the same entries.  Is there a way I can eliminate this duplication?  Thanks.
    Here is what I'm using in fstab to mount the shares:
    nfs noauto,x-systemd.automount,x-systemd.device-timeout=10,rw,rsize=32768,wsize=32768,timeo=14,hard,intr,user 0 0

    I indicated that I was trying to mount several NFS shares off my home server.  I should have mentioned that these are NOT NFSv4 shares.  Originally, I had the mount points inside my home folder.  I've since moved them in the /mnt directory and don't have duplicate entries.  Well, I don't have any entries under either Network or Devices.  I've made links to the directories where the NFS shares are mounted.  I was more curious if this was a bug with Nautilus or my own error as having the mount points in my home directory with other file browsers such as Dolphin in KDE did not yield duplicate entries with the same fstab mount options.
    I guess the long and short of it is that if you are using Nautilus, have your NFS mount points outside your home folder.
    Thanks.

  • Login problems with create pdf,share and my files

    Hello ,
    When I currently go to www.acrobat.com on my windows xp
    computer and try to go to one of the create pdf,share and my files
    and click on begin I get a message saying sorry this service is
    temporary unavailable but when I click on learn more it says all
    service are up and runing. Does anyone know what is wrong and if
    the service is down when will it be fixed.?

    Hi Sue!
    This seems like a separate issue you're running into. It is
    good that you can get logged in, but odd that you're only seeing
    partial functionality in your user.
    How many files are missing? Is this affecting certain file
    types?
    Did you ever see files in your My Files section?
    How long have they been unavailable/missing?
    When uploaded, did they complete making it through the
    scanning process?
    Can you see files in your Buzzword organizer? (which is a
    different section than your My files organizer)
    Create new documents?
    It would be good to get another browser as a comparison test
    here as well, to see if there is any different behavior!
    It could be firewall related, but what you could do is maybe
    try logging into your account from another person's computer for a
    moment, to see if the files are visible there?
    Otherwise, the test from home, or laptop at a coffee shop,
    would be another good test!
    Get back to us with some of these questions answered, and
    we'll see about getting you up and running!
    Cheers,
    Pete

  • How can a mount a NFS share exported from OpenBSD?

    Hello Apple Discussions:
    I've been experimenting with NFS in a mixed OS environment, and have been successful exporting nfs share with tigerserver, and mounting it on both a powerpc linux system, and on a powerpc openBSD system.
    Likewise, I can export a NFS share from the linux powerpc box, and mount it on the openBSD box and on the tigerserver, although, the latter required using the options (ro,sync,insecure) in my exports file.
    However, when I export a share on the OpenBSD box, I can mount it on the linux box, but not on tigerserver.
    I would like for the OpenBSD box to export a NFS share securely, with read-write permissions, to the tigerserver.
    After reading so many tutorials, that it would be a page of links, just to list them all, I am pulling my hair out. However, I have found one thread that suggests, that perhaps what I'm trying to do is impossible:
    http://www.bsdforums.org/forums/showthread.php?t=54308
    Here it is suggested that the NFS won't work because tigerserver is not using UTF-8?
    I will have to say, that I was somewhat alarmed, that the only times I succeeded in mounting an nfs share exported from linux onto tigerserver, it was when the "insecure" option is used in the /etc/exports file. There doesn't seem to be an equivalent for the linux style exports option "insecure", in the bsd style options of --maproot=user:group1:group2.
    But I don't like using any options that say "insecure" anyways, so rather than trying to find out how to make openbsd "insecure", I would rather like to find out if there is a way to get tigerserver using UTF-8, at least when mounting NFS shares, if this is indeed the issue.
    Here are the more technical details. I've created a user on all sytems named "fives" with the userid of 5555 and the groupid of 5555. I made the user local user in the local net info domain, but I've tried it with an LDAP user as well. The folders I wish to export and the folders into which to mount them are all owned by user fives and group fives, and have permissions set to 0775. The ip addresses are OpenBSD=192.168.222.111 TigerServer=192.168.222.233 LinuxPPC=192.168.222.253. I've included the relevant NFS setup files and running processes below:
    ON THE OPENBSD BOX:
    #/etc/exports
    /fives -alldirs -network=192.168.222.0 -mask=255.255.255.0
    /exports/fives -mapall=fives:fives 192.168.222.233 192.168.222.253
    #/etc/hosts.deny
    ALL: ALL
    #/etc/hosts.allow
    ALL: 192.168.222.233 192.168.222.253
    #/etc/rc.conf.local
    portmap=YES
    lockd=YES
    nfs_server=YES
    #here's proof that the daemons are running on the OpenBSD box;
    rpcinfo -p localhost
    program vers proto port
    100000 2 tcp 111 portmapper
    100000 2 udp 111 portmapper
    100005 1 udp 863 mountd
    100005 3 udp 863 mountd
    100005 1 tcp 613 mountd
    100005 3 tcp 613 mountd
    100003 2 udp 2049 nfs
    100003 3 udp 2049 nfs
    100003 2 tcp 2049 nfs
    100003 3 tcp 2049 nfs
    100021 1 udp 895 nlockmgr
    100021 3 udp 895 nlockmgr
    100021 1 tcp 706 nlockmgr
    100021 3 tcp 706 nlockmgr
    # actually, I don't see statd, but haven't found the equivalent in openbsd. There's rpc.rstatd, and maybe it should be listed here, but there doesn't seem to be a way to launch it directly. This is a competitor with the UTF-8 theory about why it's not working.
    ON THE TIGER SERVER:
    # here's proof that tiger server sees the mounts:
    showmount -e 192.168.222.111
    Exports list on 192.168.222.111:
    /fives 192.168.222.0
    /exports/fives 192.168.222.233 192.168.222.253
    # here's the result of user fives' attempt at mounting a share:
    sudo mount -t nfs 192.168.222.111:/exports/fives /imports/fives
    mount_nfs: /imports/fives: Permission denied
    # yet user fives has no problem mounting same share on linuxppc box.
    What is different about OSX server? I thought it was supposed to speak NFS?
    ---argh... I'm steppin out for a pint.. Hopefully when I'm back it'll just work.

    One thing not mentioned is that if you decide on the multiple user approach, you can have your music folder in Shared Documents so you only store the tracks once.
    Each user is free to choose which of those tracks they want in their library.
    There is an Apple help article on multiple users.
    http://docs.info.apple.com/article.html?artnum=300432

  • Custom NFS share point directory showing up on all network machines

    Hi,
    I'm in the process of migrating our 10.4 PowerMac server to a Mac Pro (running 10.5). I've been trying to recreate our 10.4 server setup as much as possible and so far I've only come across one annoying issue.
    We have fink installed on the server and under our 10.4 setup the /sw directory was set up as an NFS automounted share point with a custom mount point of '/sw'. I.e. users logging into client machines saw a /sw directory and could work with that. This made it easier to add fink packages as I only needed to do this on one machine (the server). This setup worked very well under 10.4 and had been working stably for the last couple of years.
    As we now have (for another month or two at least) a mix of intel and Power PC machines, I don't want to share out the (intel) server version of fink to all clients. In Server Admin, I have chosen to set the NFS protocol options to specify the IP address of just one client (an intel machine). I am only using NFS to share this directory. The plan is to add more client IP addresses as we get more intel machines.
    This works for the one intel client machine. Logging in via the GUI or via ssh allows you to run programs located under the /sw directory. The problem is that a phantom /sw directory appears on all client machines, even though their IP addresses are not specified in Server Admin. The /sw directory has root/wheel permissions (for user/group) and attempting to list its contents returns 'Operation not permitted' (even with 'sudo ls /sw').
    If I use Directory Utility to remove the connection to the Directory server on our main server, then the /sw directory becomes owned by root/admin and I can remove it (it appears empty). Reconnecting to the Directory Server changes the permissions back to root/wheel. It is also worth noting that when I first installed fink on the server (in /sw) the act of making this a share point also changed the permissions on /sw to root/wheel meaning that I couldn't access the fink programs that I had only just installed (this forced me to reinstall fink in /Volumes/Data/fink).
    Has anyone else noticed this behavior? It almost seems like Server Admin is not honoring the list of IP addresses that are being listed as targets for client machines. I had planned to install fink locally on the PowerPC clients until we upgrade them to intel machines. However, I would then also have to install fink somewhere other than /sw as I can't write to that directory. I would presume that this behavior should happen on any NFS share point that is trying to automount to a custom mount point on a client. Can anyone else verify this?
    Regards,
    Keith

    As a footnote. I have now removed my shared fink installation. It is no longer listed as an NFS sharepoint in Server Admin and running the 'showmount -e' command does not list it. However, a /sw directory is still being created on the server and on the client machines on our network.
    This is perplexing and frustrating. There is no sharepoint any more. I rebooted the server but it makes no difference. I removed the /sw directory (on the server) by booting the machine in target firewire mode and removing it by using a 2nd machine. But following the restart, it appeared again.
    This suggests that once you make any custom mountpoint (using NFS sharing) then you will forever be stuck with a directory at the root level of all your clients which you can not remove.
    Keith

  • Launching xcode from nfs share (Ensure that Xcode.app is installed on a volume with ownership enabled)

    Hi!
    We have mac mini (Yosemite) and NFS server under Ubnuntu 14.04.
    Also we have xcode resides on NFS share that mounted to mac.
    Problem: When I try launch xcode from nfs share I got error message:
    NSLocalizedRecoverySuggestion=Ensure that Xcode.app is installed on a volume with ownership enabled
    You can see full error here:
    https://gist.github.com/keferoff/fcfd3ea6c13f6ba481fa
    The question is:
    How I can launch xcode from nfs share? For some reasons I can't use xcode-select or store several xcodes locally.
    Thanks in advance!

    It's cool answer but I need solution how to accomplish my task. Maybe I can use iSCSI device or NBD device ot maybe there we have some NFS mount options?

Maybe you are looking for