Wrong atime created by iMac as a NFS client

We have iMac computers in classrooms, and they are running as an NFS client. We have found that the wrong access time (atime) is created by Mac OS 10.5 and 10.6.
A file that is opened by using O_CREATE on the NFS-mounted directory will have the wrong atime. For example, the year will appear as 1920 or 2037 depending on the IP address of the NFS client.
This problem can also be found with Solaris NFS server and EMC Celerra.
The following program will create a file having the wrong atime on an NFS-mounted directory:
#include <fcntl.h>
#include <stdlib.h>
#include <string.h>
#include <stddef.h>
main() {
int fd;
/* O_EXCL option will cause wrong atime */
fd = open("AIZU", O_CREAT | O_EXCL | O_RDWR, 0644);
/* fd = open("AIZU", O_CREAT | O_RDWR, 0644); */
close(fd);
exit(0);
The phenomenon is quite similar to the following problems on FreeBSD reported in 2001:
http://www.mail-arch...g/msg22084.html
If anyone knows a solution to this problem, we would greatly appreciate your advice. Thanks in advance.

The solution shown in
http://www.mail-archive.com/[email protected]/msg22084.html
should be applied to the current NFS modules of Snow Leopard.

Similar Messages

  • HT204053 I have a problem. I bought Pages and Numbers for my iMac and so far everything was OK.  Suddenly i have a problem with connection to iCloud. Dokuments created on iMac in Pages or Numbers are not transfered to iCloud. Not uploaded in iCloud.

    I have a problem.
    I bought Pages and Numbers for my iMac and so far everything was OK.
    Suddenly i have a problem with connection to iCloud.
    Dokuments created on iMac in Pages or Numbers are not transfered to iCloud. Not uploaded to iCloud. There is a sign  (small shape of iCloud), that they are waiting to be uploaded to iCloud.
    I think the iCloud is OK.
    i can see and manage documents in iCloud. i can create them on iPad and  see them and delete them in iCloud.
    But not on iMac.
    i dont know what has happened and I am not abble to find the solution. It was OK till some days ago….

    I have a similar problem! Everything stopped working after upgrading to Mac OS 10.8.4 through the App Store. Help!

  • Have a problem with my MacBook Pro.5,3 and monitor.  Intermittently, the desk top screen becomes broken up into vertical sections, which appear in the wrong order, creating a haphazard picture.    The Apple icon, usually at top left corner, appears in to

    have a problem with my MacBook Pro.5,3 and monitor.  Intermittently, the desk top screen becomes broken up into vertical sections, which appear in the wrong order, creating a haphazard picture.
    The Apple icon, usually at top left corner, appears in top right corner  in a strip including all that left hand.  When the computer is restarted, the screen goes back to normal.  But I have to move the mouse to the far left of the screen and then up until it reappears on the far right hand side of the screen so that I can hit shut down button.
    It seems to happen randomly though always on desk top and sometimes when I move mouse and computer switches on from rest mode.   My large flat screen is relatively new.
    The video card on this computer, which is approximately two years old, is Nvidia FeForce 9400M.
    Any help would be welcome as my usual source of assistance is not available and I am not well versed in workings of the Mac although I have used one for many years.

    That sounds like a Hardware problem.
    You may be able to confirm that with Apple Hardware Test, or make an appointmenst for a free Genius Bar evaluation.

  • Bridge CS6 occasionally displaying wrong 'Date Created' for .NEF's

    I noticed this problem only recently, and this the third time it happens (that I know of).
    Sometimes my Bridge puts a wrong Date Created (not Date File Created or Date File Modified) on a NEF RAW file, whilst the next one (taken perhaps just a few seconds or a minute later) has the proper date. Strangely the Date File Created & Modified do show the proper creation date.
    Example to clarify:
    Metadata viewed on my D90 itself
    Photo 1 taken on: 30/09/2013, 9:03:26 PM (DD-MM-YYYY)
    Photo 2: 30/09/2013, 9:07:52 PM
    Photo 3: 30/09/2013, 9:08:12 PM
    Photo 4: 30/09/2013, 9:08:34 PM
    Photo 5: 30/09/2013, 9:08:54 PM
    Photo 6: 30/09/2013, 9:09:34 PM
    viewing in bridge:
    Photo 1: Date Created: 30/09/2013, 9:03:26 PM
                   Date File Created: 30/09/2013, 9:03:26 PM
                   Date File Modified: 30/09/2013, 9:03:26 PM
    Photo 2: Date Created: 30/09/2013, 9:07:52 PM
                   Date File Created: 30/09/2013, 9:07:52 PM
                   Date File Modified: 30/09/2013, 9:07:52 PM
    Photo 3: Date Created: 23/09/2013, 7:42:02 PM
                   Date File Created: 30/09/2013, 9:08:12 PM
                   Date File Modified: 30/09/2013, 9:08:12 PM
    Photo 4: Date Created: 30/09/2013, 9:08:34 PM
                   Date File Created: 30/09/2013, 9:08:34 PM
                   Date File Modified: 30/09/2013, 9:08:34 PM
    Photo 5: Date Created: 04/09/2013, 10:14:45 AM
                   Date File Created: 30/09/2013, 9:08:54 PM
                   Date File Modified: 30/09/2013, 9:08:54 PM
    Photo 6: Date Created: 04/09/2013, 10:15:06 AM
                   Date File Created: 30/09/2013, 9:09:34 PM
                   Date File Modified: 30/09/2013, 9:09:34 PM
    As you can see, it's not a 'stable' deviation.
    I could live with it, apart from one problem: the batch rename function only has 'Date Created' as option for renaming with date.
    It's also quite annoying for my photography study, since my teachers might get confused by the wrong date in the Metadata when I turn my photo in.
    I tried compacting the cache whilst my camera's SD card wasn't connected (so the files aren't 'available' anymore), no difference.
    Another thing I tried was to hook my SD card up on a different computer, same result.
    So it might be my SD card, but that doesn't explain why the correct date is shown on my D90 itself, though.
    If anybody here has an idea how to fix / avoid this problem, help is much appreciated.
    Greetings,
    Alex

    I have never inserted the 8GB card in the D3000 before just a minute ago.
    And I just took 7 rapid photo's of my previous post (~1s interval)
    All 7 displayed actual date in my camera, the first 4 displayed a wrong date in bridge (24/09/2013, times; 2:12:16PM, 2:12:20PM, 3:09:24PM and 3:10:26PM). Last 3 displayed proper time in bridge.
    So yea, I'm pretty sure something is wrong. I'll try formatting the SD card (as suggested before) but however you look at it, this doesn't add up.
    edit:
    Well I figured out one thing: if I view the photo's directly on my SD card through bridge, it shows the same date for Date Created, Date File Created and Date File Modified.
    Quite logical, to be honest, since I am viewing the date the file was created on the camera.
    However, if I export photo's with Adobe Photo Downloader (or whatever it's called again), it displays the date those copies were created (by Date File Created). Date File Modified is still ~the same as Date Created since it hasn't been modified since the original creation.
    edit2:
    The formatting seems to have solved the issue, for now anyway. Still have no clue why the SD card got such (occasional) cases of drunkness.

  • I'd copied some pictures from a PC to a folder in Finder in my new iMac. The 'Date Created' in iMac is not the same as the 'Date Created' in the PC. Why? Is there a way to fix that?

    I'd copied some pictures from a PC to a folder in Finder in my new iMac. The 'Date Created' in iMac is not the same as the 'Date Created' in the PC. Why? Is there a way to fix that?

    View Menu -> Sort Photos is a good way to start.
    Some organising possibilities in iPhoto:
    I use Events simply as big buckets of Photos: Spring 08, July - Nov 06 are typical Events in my Library. I use keywords and Smart Albums extensively. I title the pics broadly.
    I keyword on a
    Who
    What
    Where basis (The When is in the photos's Exif metadata). I also rate the pics on a 1 - 5 star basis.
    Using this system I can find pretty much find any pic in my 50k library in a couple of seconds.
    So, for example, I have a batch of pics titled 'Seattle 08' and a  typical keywording might include: John, Anne, Landscape, mountain, trees, snow. With a rating included it's so very easy to find the best pics we took at Mount Rainier.
    File -> New Smart Album
    set it to 'All"
    title contains Seattle
    keyword is mountain
    keyword is snow
    rating is 5 stars
    Or, want a chronological album of John from birth to today?
    New Smart Album
    Keyword is John
    Set the View options to Sort By Date Ascending
    Want only the best pics?
    add Rating is greater than 4 stars
    The best thing about this system is that it's dynamic. If I add 50 more pics of John  to the Library tomorrow, as I keyword and rate them they are added to the Smart Album.
    In the end, organisation is about finding the pics. The point is to make locating that pic or batch of pics findable fast. This system works for me.

  • Can a slide show created on iMac be shown on an iPad ?

    can a slide show created on iMac be shown on an iPad ?

    If you use keynote on both, yes.
    You can also have keynote remote on your iphone to control the slides on your ipad.

  • NFS client problem "The document X could not be saved"

    Hi,
    Briefly: Debian Linux server (Lenny), OS X 10.5.7 client. NFS server config is simple enough:
    /global 192.168.72.0/255.255.255.0(rw,rootsquash,sync,insecure,no_subtreecheck)
    This works well without our Linux clients, and generally it is Ok with my OS X iMac. OS X NFS client is configured through Directory Utility, with no "Advanced" options. Client can authenticate with NIS nicely, and NFS, on the whole, works. I can manipulate files with Finder, and create files on the commandline with the usual tools.
    The problem is TextEdit, iWork and other Cocoa apps (not all). They can save a file once, but subsequently saving a file produces a "The document X.txt cannot be saved" error dialog. If I remove the file on the commandline and re-save, then the save succeeds. It is as if re-saving the document with the same name as an existing file causes issues. There seems to be no problem with file permissions. When I save in a non NFS exported directory everything is fine.
    Has anyone spotted this problem before?
    Lawrence

    I doubt that "OS X NFS is fundamentally broken" seeing as how many people use it successfully.
    tcpdump (or more preferably: wireshark) might be useful in tracking down what's happening between the NFS client and NFS server. Sometimes utilities like fs_usage can be useful in tracking down the application/filesystem interaction.
    It's usually a good idea to check the logs (e.g. /var/log/system.log) for possible clues in case an error/warning is getting logged around the same time as the failure. And if you can't reproduce the problem from the command line, then that can be a good indication that the issue is with the higher layers of the system.
    Oh, and if you think there's a bug in the OS, it never hurts to officially tell Apple that via a bug report:
    http://developer.apple.com/bugreporter/
    Even if it isn't a bug, they should still be able to work with you to help you figure out what's going on. They'll likely want know details about what exactly isn't working and will probably ask for things like a tcpdump capture file and/or an fs_usage trace.
    HTH
    --macko

  • NFS client test program

    I'm trying to write a test program that behaves like an NFS client. The code uses the standard RPC calls to talk to the NFS server. It compiles and runs fine on Linux. On Solaris though I have a problem establishing an RPC client initially.
    The test program performs the following operations:
    1. Uses the portmapper service on the server to establish the remote port associated with the MOUNT service
    1. Creates a socket, sockfd.
    2. Uses bindresvport to bind a reserved port to the socket. The port number is in the reserved range (ie 600 to 1000ish).
    3. Connects this socket to the server
    3. Creates a struct sockadd_in with the appropriate address information for the server in addr
    4. Calls clnttcp_create
    The following code
    client = clnttcp_create(addr, <------- points at the server
    program, <-------- MOUNT RPC program number
    program_version, <--------- MOUNT RPC program version
    &sockfd, <------- connected socket using a reserved port on the local client
    0,
    0);
    if (client == NULL) {
    clnt_pcreateerror("clnttcp_create");
    returns "clnttcp_create: RPC: Remote system error - Address already in use".
    So it looks like port 688 (the local port selected say by bindresvport) is already in use ? But that's the whole point, I need to provide cnttcp_create() with a socket that already has a port bound to it.
    The same code runs fines on a Suse 11 client talking to Open Solaris 11 server. The same code run as a client on Solaris 11 fails with the above message when talking to either a Suse 11 server or a Solaris 11 server.
    Any suggestions what I'm doing wrong? Is this the right forum in which to ask?
    many thanks

    Make sure you have the sub-folder com, and then another sub-folder underneath it, mastertech and another sub-folder underneath this called sample. This is how packages are structured. They should be in your ep workspace directory structure if you're using Eclipse. In JBoss, you do not have to have the packages or any source code structure. Just jar your compiled code and save in JBoss's deploy folder which only accepts jar's, war's and ear's.

  • Systemd nfs client mount share ???

    Just installed systemd, systemd-arch-units and initscripts-systemd  as per the wiki and all went well except mounting a nfs share from my server. (networking is OK!)
    This is the old working mount command run from rc.local before installing systemd;
    mount -t nfs4 -o rw,hard,async,intr,rsize=49152,wsize=49152,proto=tcp 192.168.0.250:/ /media/SERVER_NYTT &
    Did not mount at all with systemd, not even when run manually from command line. From systemd log netfs and nfs-common failed to start.
    So I tried from fstab instead;
    192.168.0.250:/ /media/SERVER_NYTT nfs rw,hard,async,intr,rsize=49152,wsize=49152,proto=tcp 0 0
    This did work but 'mount' showed that systemd had mounted with nfs default options instead of mine (wsize, rsize ...). Still errors from systemd starting netfs and nfs.
    So I disabled netfs and nfs-common/rpcbind from rc.conf and created this systemd file (/etc/systemd/system/192.168.0.250.mount);
    [Unit]
    Description=ServerNfs
    Wants=network.target rpc-statd.service
    After=network.target rpc-statd.service
    [Mount]
    What=192.168.0.250:/
    Where=/media/SERVET_NYTT
    Type=nfs
    Options=rw,hard,async,intr,rsize=49152,wsize=49152,proto=tcp
    DirectoryMode=0777
    StandardOutput=syslog
    StandardError=syslog
    From the sparce wiki and 'man systemd.mount'. Now nothing happens. With my limited understanding I thought it would start neccessary services (wants) and replace the entry in fstab.
    I will now enable systemd services rpcbind, rpcstatd and see what happens.
    Overall the transition to systemd went very well indeed; slim, openbox, network, .xinitrc, e4rat all started OK much to my surprise! There's still some fine-tuning to do like nfs, and possibly automounting and at last weeding out unneccessary services.
    But any help with this nfs client problem is much appreciated.

    swanson wrote:
    Do you use any of this;
    Alternatively, you can mark these entries in /etc/fstab with the x-systemd.automount and x-systemd-device-timeout= options (see systemd.mount ..
    I did read the man for systemd.mount but couldn't make out what to put in the fstab line.
    By the way, I renamed the mountfile to the target mountpoint but no success. I then reactivated the fstab line, with the new mountpoint in /mnt and that worked fine with mount -a. Except the mount options are not as I want, as they are in fstab. They were before systemd, and now they are as default for nfs4.
    have a look here regarding the x-systemd-automount option. I seem to be only one who has noticed a problem with this option, comment=systemd.automount works fine for automounting though. I haven't changed the wiki because I'm still not sure if the problem is on my side.

  • Is NFS client data cacheing possible?

    Yesterday, I was viewing an HD 1080 video with VLC, and noticed that Activity Monitor was showing about 34MB/sec from my NAS box. My NAS box runs OpenSolaris (I was at Sun for over 20 years, and some habits die hard), and the 6GB video file was mounted on my iMac 27" (10.7.2) using NFSv3 (yes, I have a gigabit network).
    Being a long term UNIX performance expert and regular DTrace user, I was able to confirm that VLC on Lion was reading the file at about 1.8MB/sec, and that the NFS server was being hit at 34MB/sec. Further investigation showed that the NFS client (Lion) was requesting each 32KB block 20 times!
    Note: the default read size for NFSv3 over TCP is 32KB).
    Digging deeper, I found that VLC was reading the file in 1786 byte blocks. I have concluded that Lion's NFSv3 client implement at least one 32KB read for each application call to read(2), and that no data is cached betweem reads (this fully accounts for the 20x overhead in this case).
    A workaround is to use say rsize=1024, which will increase the number of NFS ops but dramatically reduce the bandwidth consumption (which means I might yet be able to watch HD video over wifi).
    That VLC should start issuing such small reads is a bug, so I have also written some notes in the vlc.org forums. But client side cacheing would hide the issue from the network.
    So, the big question: is it possible to enable NFS client data cacheing in Lion?

    The problem solved itself mysteriously overnight, without any interference from myself.
    The systems are again perfectly happily mounting the file space (650 clients of them all at the same time
    mounting up to 6 filesystems from the same server) and the server is happily serving again as it has been for the past 2 years.
    My idea is that there has been a network configuration clash, but considering that the last modification of NIS hosts file was about 4 weeks ago and the latest server was installed then and has been serving since then, I have no
    idea how such clash could happen without interference in the config files. It is a mystery and I will have to make
    every effort to unravel it. One does not really like to sweep incidents like that un-investigated under the carpet.
    If anybody has any suggestions and thoughts on this matter please post them here.
    Lydia

  • I can not create an email account for my client

    I have created an email account ([email protected]) and it is sending snd recieving messages. However when I try to create an address ([email protected]) me client could not log in to the webmail. I deleted the account and re-created the address with my email as the recovery. My thinking was I would set up the email and confirm everything is working and then show the client how to reset the password. Now it I can not get the BC admin panle to send an invite email yo me. Please advise. Client is passed mad on this. I have to resolve this issue.

    When I try to set up the [email protected] email account in the BC admin panle it repeatedly continues to change the password recovery email to [email protected] no matter how I set it up. This doesn't make any sense.  We can get into that account to even set a user password.

  • Svc:/network/nfs/client problem in Solaris 10

    Hi,
    I`ve been trying to figure out what is the problem with the below scenario for so long and till now i`m not able to.
    i have a X4100 SunFire system with solaris 10 installed.
    most of the services are not coming up (ssh,ftp,etc....) each time i reboot the server, and after approximately 3 hours everything comes fine by it self.
    i noticed that the service :/network/nfs/client is taking too long while starting. This might be related to the problem.
    See below the svcs command : hope this will be usefull :)
    bash-3.00# svcs -xv ssh
    svc:/network/ssh:default (SSH server)
    State: offline since Wed May 20 16:37:45 2009
    Reason: Service svc:/network/nfs/client:default is starting.
    See: http://sun.com/msg/SMF-8000-GE
    Path: svc:/network/ssh:default
    svc:/system/filesystem/autofs:default
    svc:/network/nfs/client:default
    See: man -M /usr/share/man -s 1M sshd
    Impact: 3 dependent services are not running:
    svc:/milestone/multi-user-server:default
    svc:/system/basicreg:default
    svc:/system/zones:default
    any advise on this ?
    Appreciate your help,

    hi all
    issue resolved actually port was conflicting in /etc/services i have readded the entry in /etc/service
    Thanks

  • How to create a Credit Note to the client for purchasing of an Asset last y

    Hi,
    I have an issue in Asset Accounting as it showed bellow:
    Assets depreciation start date is 1st October 2007.
    We have run the depreciation up to November 2007.
    Than we sold the asset in December.
    However, after the end closed year 2007, client returned the asset in 1 st January 2008.
    So i need to create a Credit Note to the client for purchasing of an Asset last year, but we don't know how to create it?
    We thinks that correct movement should be debit in extraordinary cost / profits account, debit in asset account, credit in accumulated depreciation account and credit in client account.
    Plz help. I will assign full points for any useful answer.
    Thanks
    Edited by: Ricardo Pedro Rodrigues Ferrão on Dec 20, 2007 10:09 AM

    How did you sell the asset?
    You can just reverse the sale transaction by going to transaction AB08.
    Hopefull this may work for you.
    Regards,

  • I have created a Muse site for a client that wishes to host with Business Catalyst. How do I publish the site with their account rather than using one of my free sites?

    I have created a Muse site for a client that wishes to host with Business Catalyst. How do I publish the site with their account rather than using one of my free sites? This is so I can keep my free ones for personal projects but also so they can pay for their own hosting. I am happy to set it all up for them but not sure what to do.

    Hi
    You can use their BC login details and use them , which on publish the site will be under their account.
    Please change the BC login from Edit > Preferences > Publish > Switch Accounts , for Mac it would Adobe Muse > Preferences
    Thanks,
    Sanjit

  • [Solved]NFS Client Not Mounting Shares

    Here is my setup:
    I have two Arch boxes that I am attempting to setup NFS shares on.  The box that is going to be the server is headless FYI.  So far, I have installed nfs-utils, started `rpc-idmapd` and `rpc-mountd` successfully on the server, and started `rpc-gssd` successfully on the client.
    The folder I am trying to share is the /exports folder.
    ls -l /exports
    produces
    total 8
    drwxrwxrw-+ 110 daniel 1004 4096 Dec 6 17:26 Movies
    drwxrwxrwx+ 13 daniel users 4096 Jan 8 19:12 TV-Shows
    On the server:
    /etc/exports
    # /etc/exports
    # List of directories exported to NFS clients. See exports(5).
    # Use exportfs -arv to reread.
    # Example for NFSv2 and NFSv3:
    # /srv/home hostname1(rw,sync) hostname2(ro,sync)
    # Example for NFSv4:
    # /srv/nfs4 hostname1(rw,sync,fsid=0)
    # /srv/nfs4/home hostname1(rw,sync,nohide)
    # Using Kerberos and integrity checking:
    # /srv/nfs4 gss/krb5i(rw,sync,fsid=0,crossmnt)
    # /srv/nfs4/home gss/krb5i(rw,sync,nohide)
    /exports 192.168.1.10(rw,fsid=0)
    On the client:
    showmount -e 192.168.1.91
    Export list for 192.168.1.91:
    /exports 192.168.1.10
    Everything is looking hunky-dory.  However, I go to mount using
    sudo mount -t nfs4 192.168.1.91:/exports /mnt/Media
    and the mount never takes place.  It sits there and does nothing.  I CAN, however, kill the process with Ctrl-c.
    So does anybody have ANY idea why my shares aren't working.
    EDIT: Just thought I should mention that all of the data in the /exports folder is a mount --bind from /mnt/media.  All of the /mnt/media is contained on a USB external hard drive.  I did notice that there is an ACL.
    getfacl /exports
    getfacl: Removing leading '/' from absolute path names
    # file: exports/
    # owner: root
    # group: root
    user::rwx
    group::r-x
    other::r-x
    Last edited by DaBungalow (2014-01-10 03:18:05)

    I found what the problem was.  Apparently rpc_gssd was causing a problem.  Stopping it fixed everything.

Maybe you are looking for