Automator copy files incrementally by date?

I need to daily backup a folder from my mac (4gb folder) to a windows server. I can't use time machine, since the hard drive on the server is formated differently then what it needs to be to use time machine.
I've been using automator's copy finder file, but it just copies the whole thing, and 4gb takes too long
I've seen people scripting to do such copies while ignoring files that already have the same name. But that doesn't help me because I need to copy all files that have been modified, regardless of their name. Maybe someone can help me write an apple script that copies files if their "last modified" dates are different?
Or maybe there is a solution to using time machine with a windows-formated folder?
Thanks ;P

Short answer: No. That is the way it is.
Why? The Make Second Copy during Import is designed to be a backup of your memory card processes-not a back up of your files as you store them on disk. It is intended to be a transient copy that allows you an immediate backup until such time as a File System Level backup can be performed, allowing you to have two copies so that you can reformat your memory cards.
I agree it would be nice if the folder storage strategy was duplicated for the backup as it is created for the import but that is not something you can set the program to do currently.  I remember it behaving differently versions ago but do not recall exactly how it used to work.

Similar Messages

  • Copy File name and Date only from a Folder with SSIS

    Hello all,
    I need to create a package which will list the file names and the dates in a folder, these files are blank flat files. After extracting the file names from the file path i need to create a table which will have two columns Filename and
    Date.We get about 1000 such files every month and we are looking to maintain them in a table.
    Is this possible? Can someone point me in the right direction? 
    FM

    Arthur , I actually figured it out. The For each container let me pass thru the file name, then I inserted the file name into my staging table. I created a FilePath variable and basically Did a sql task and
    INSERT INTO MyTable
    SELECT Filename = ?,
    LoadDate = GetDate()
    In my for each loop I added a *.*txt qualifier... I`ve tested and works well... Got what I needed.
    FM

  • Can I copy files – but with certain restrictions?

    I have just finished a large project, the archives of which involves about 5000 "base" files, stored on about 80 CDs and 50 DVDs, involving about 50,000 files in total. Each of the "base" files may have had up to 30 incremental versions. i.e. a certain text file may have undergone revision 23 times, and each revision was saved and archived to (probably) a different disk, with a different suffix – a, b, c and so on. But sometimes the suffix didn't change even though the file was edited. I might have done a bit more dust removal on an image and just overwrote the old file (already archived), and so the new one was archived on a different disk.
    I now have 130 disks from which I would like to extract all the files and collapse them to one large archive that will probably span about 20 disks by the time I delete some files not needed. That way I can easily search for all versions of, say, GB097, by going to the particular DVD that has the "G" files on it. Up would come:
    GB097
    GB097a
    GB097b
    GB097b-1
    GB097b-2
    GB097c
    ... and so on.
    This is what I would like to do:
    1. Grab the first archive disk, open every folder, and copy all the files to the one folder on a hard drive.
    2. Open the second disk and repeat step (1), but with these two provisos.
    (a) If a file is identical to a previously copied file (maybe I archived it twice), the file isn't copied. However...
    (b) If a file has the same name as a previously copied file, but the data within that file is different (i.e. I removed some dust from an image file, but left the name unchanged), I'd like that file to be copied with a numbered suffix, the same way that Trash treats identically named files.
    Any suggestions how I could do this?

    Back again. This is the Terminal output when run on my +Duplicates Copy+ folder:
    Last login: Thu May 20 13:28:30 on ttyp1
    Welcome to Darwin!
    jenny-pearces-imac-g5:~ Jenny$ chmod +x /Users/Jenny/Desktop/Bash\ Test\ Documents/BashTest.txt
    jenny-pearces-imac-g5:~ Jenny$ /Users/Jenny/Desktop/Bash\ Test\ Documents/BashTest.txt /Users/Jenny/Desktop/Duplicates\ copy/
    /Users/Jenny/Desktop/Bash Test Documents/BashTest.txt: line 1: ${filePaths${count}}: bad substitution
    /Users/Jenny/Desktop/Bash Test Documents/BashTest.txt: line 17: fileSum0=d41d8cd98f00b204e9800998ecf8427e: command not found
    /Users/Jenny/Desktop/Bash Test Documents/BashTest.txt: line 1: ${filePaths${count}}: bad substitution
    /Users/Jenny/Desktop/Bash Test Documents/BashTest.txt: line 17: fileSum1=d41d8cd98f00b204e9800998ecf8427e: command not found
    /Users/Jenny/Desktop/Bash Test Documents/BashTest.txt: line 1: ${filePaths${count}}: bad substitution
    /Users/Jenny/Desktop/Bash Test Documents/BashTest.txt: line 17: fileSum2=d41d8cd98f00b204e9800998ecf8427e: command not found
    jenny-pearces-imac-g5:~ Jenny$
    I'm not sure it's worth spending much more time on this. However, if you're prepared to keep posting, JJJ, I'll keep testing.
    *SCRIPT AFTER THE EDIT*
    #!/bin/bash
    declare -i count=0
    declare -a fileSum filePaths
    # Main folder can be declared on the command line or if not then use /testpix
    MainFolder="${1:-/testpix}"
    # Set the internal field separator to newline to preserve spaces in file paths
    IFS=$'\n'
    # Use 'find' to create a list of files within the folders.
    filePaths=( $( find "${MainFolder}" -type f \! -name ".*" ) )
    # Get an MD5 checksum for each file's combined content of both data and resource forks
    for file in ${filePaths[*]} ; do
    fileSum${count}=$( cat "${filePaths${count}}" "${filePaths${count}}/rsrc" | md5 | cut -d'=' -f 2)
    let count+=1
    done
    # For each file, check for a duplicate checksum and if found, move the matching file to the user's trash folder
    # Rename files with duplicate names by appending #XX
    for ((i=0;i<${count};i++)) ; do
    [ -z "${filePaths[${i}]}" ] && continue
    dupecount=1
    for ((j=0;j<${count};j++)) ; do
    [ -z "${filePaths[${j}]}" ] && continue
    if [ "${fileSum[${i}]}" = "${fileSum[${j}]}" -a ${i} -ne ${j} ] ; then
    mv "${filePaths[${j}]}" ~/.Trash && filePaths[${j}]='' && fileSum[${j}]=''
    elif [ $(basename "${filePaths[${i}]}") = $(basename "${filePaths[${j}]}") -a ${i} -ne ${j} ] ; then
    let dupecount+=1
    dirname=$(dirname "${filePaths[${j}]}")
    fullfilename=$(basename "${filePaths[${i}]}")
    extension="${fullfilename##*.}"
    filename="${fullfilename%.*}"
    until [ ! -e "${dirname}/${filename} #${dupecount}.${extension}" ] ; do
    let dupecount+=1
    done
    mv "${filePaths[${j}]}" "${dirname}/${filename} #${dupecount}.${extension}"
    filePaths[${j}]="${dirname}/${filename} #${dupecount}.${extension}"
    fi
    done
    done
    exit 0

  • Copied Files From CD/DVD Locked

    Hiya folks,
    When i copy files from a data cd/dvd to my home partition they are locked and can only be modified by root. How do I gain permission?
    Thanks
    Andy

    Ghost1227 wrote:if only root can edit a file stands to reason that only root can change permissions on it!
    I would say a user can change rwx-permissions without su/sudo as long as (s)he is the owner of the file(s).
    I just tried it. Copying a file from cdrom wrote the file with these rights: -r-xr-xr-x and my useraccount as the owner. So I can use 'chmod +w' without su/sudo to make the file writeable.

  • Help - "Cannot copy file: Data error (Cyclic Redundancy Check)"

    Hey there,
    I'm having problems with my Zen Mosaic EZ00. Up until a few weeks ago it has been working fine however, since then, has begun to show the error message? "Cannot copy file: Data error (Cyclic Redundancy Check)" whenever I try to copy files on to it. I have tried the recovery software available and it only updated the firmware. Some files do occasionally (and very occasionally) copy but usually it shows the error message and cancels the copy.
    I would really appreciate if anyone could help me with this?
    Many thanks :smileytongue:

    sorry for the double post!
    ive just realised that my ipod isnt actually charging, its just got the orange light blinking!
    not too sure what is going on now!
    lol

  • Detection Method of Date Modified fails due to copied files date modified change

    I've been working through an issue for some time now and I think it's a bug with SCCM 2012 SP1. I've built literally 50 deployments in "Application Management" in SCCM 2012, but I've been chasing my tail on a few of them throwing the
    "succeeded but couldn't detect" error. The few that have given me problems are all copying text files and then detect the modified date to verify that the file is there. All worked when I initially built them in Windows 8 but are now failing
    randomly...even on a machine where they previously worked and I deleted the files to test the deployment again. 
    What I'm doing:
    My java deployment requires policies which we control with the Deployment.config, deployment.properties and java.policy files. A simple script creates the directories and copies the files. I then create a deployment type for this "ConfigFileCopy.cmd"
    and create a detection rule for each file that I am moving over. The detection rule looks for a "date modified" date between the beginning and end of the date for the day I modified the file. This allows me to modify the files to change our
    java policies and then redeploy them with the "new" modified date as the detection rule, which forces the update.
    The problem:
    When I first created my deployment, everything worked pretty well. The files copy out and the script runs successfully. There were some minor tweaks I need to make to my scripts, so I did and then updated the content for the deployment type. I believe this
    is where things go a little wonky. At this point, my deployment starts failing with the "0x87D00324" error.
    Looking through the logs, everything looks good. The reason configuration manager is failing on detecting the modified date for those files, as it turns out, is because the modified dates don't meet the criteria. So that part is properly failing the detection.
    The problem is that the modified dates are correct on the server but not the client. Looking in c:\windows\ccmcache, I can see multiple folders....presumably one for each version of the application that I updated. Looking at any of the newer content folders
    for this app, the modified dates are the date and time the file was copied out to the workstation, which is incorrect as the file hasn't been modified during that process.
    The odd part, is that this doesn't happen EVERY time on EVERY machine. My primary desktop is windows 8.1 and received the files correctly and installed without issue. My test win 8.1 laptop initially received them correctly, but then as I refined the scripts
    it began picking up the wrong modified dates and started failing.
    I found a similar issue to this existed in SCCM 2007 (http://support.microsoft.com/kb/2276865) so I suspect that this is truly just a bug that hasn't been addressed (or maybe is fixed in R2). Unfortunately
    we have business reasons we can't upgrade to R2 at this time so I'm hoping someone has experienced this and has some sort of work around that might get me by for now. If someone can confirm it is fixed in R2, that would help my case to upgrade as well.  
    I can work around the issue by changing my logic to detect any date greater than a specific date. But I shouldn't have to do that and I'm concerned there are scenarios I haven't thought of that will cause unexpected behavior or failures with that.
     

    This outlines the behavior. It's apparently by design. 
    "SCCM has a habit of changing the ‘Date Modified’ timestamp of all files it delivers when it detects an ‘upgrade’ of the source files for that application. It typically does not touch the timestamp
    of the source files if it delivers a brand new install to a client that has never received the software, however if a single file in the source folder is changed for that application, then SCCM tries to use a previous version of the application in the cache
    (C:\windows\ccmcache) and only downloads the new file that has change. This results in all files having their ‘Data Modified’ timestamp changing (except for the brand new file). Therefore determining if that application has delivered
    successfully using ‘Date Modified’ timestamps is not recommended. The key to seeing this process in action is looking at the file properties in the
    C:\windows\ccmcache\<sccm code> folder for that application, particularly before and after a file is updated in the original source SCCM application folder."
    http://blog.kloud.com.au/tag/sccm/

  • 2006 iMac running 10.4 OS. I copied files to ext. drive. Put in 10.6 disc and did a clean install (format) of 10.6. Copied data from ext. drive back to Mac. I didn't get all the files. CAN I REVERT BACK TO THE 10.4 OS SO I CAN COPY MY DATA PROPERLY?

    2006 iMac running 10.4 OS. I copied files to ext. drive. Put in 10.6 disc and did a clean install (format) of 10.6. Copied data from ext. drive back to Mac. I didn't get all the files. CAN I REVERT BACK TO THE 10.4 OS SO I CAN COPY MY DATA PROPERLY?

    How did you copy files to the external drive, and what files did you copy? Did you drag and drop in the Finder? Use a clone utility? What files did not successfully copy?
    If you reformatted the drive as part of your install, then you won't be able to recover any files from the drive. Reverting back to (reinstalling) 10.4 will not help you.
    Do you have a complete-drive backup of your old system?
    Matt

  • I was backing up my iphone by changing the location of library beacause i don't have enough space.My phone was taking too long to copying file so i can celled it.the data is stored in desired location . And now i can't delete that back up

    I was backing up my iphone by changing the location of library because i don't have enough space.My phone was taking too long to copying file so i can celled it.the data is stored in desired location . And now i can't delete that back up.
    Also tell me about the performance of iphone 4 with ios 7.1.1...........
    T0X1C

    rabidrabbit wrote:
    Can I back up my iPhone 4S to my ipad 3 (64 gb)?
    no
    rabidrabbit wrote:
    However, now I don't have enough space in iCloud to backup either device. Why not?
    iCloud only give so much space for free storage, then if you exceed the limit of 5gb you have to pay for additional storage.

  • Empty lines getting inserted while copying File data as chunks

    When I try to copy data from one file to other, Extra lines are getting inserted into the copied data.
    I am actually reading chunks of data from the first file and writing it to the other by using 'Byte' count to read.
    Whenever the no. of bytes read ends with the EOL, one extra line is getting inserted in the second file. 
    Consider the code snippet given below :
    If the first file content is like this :
    Then the copied file looks :
    The two files are also attached with this post.
    Please shre your suggestions to solve this issue
    Attachments:
    File to be copied.txt ‏1 KB
    Copied File.txt ‏1 KB

    Right click on your Read and Write text file functions and uncheck the Convert EOL option.
    You should also set the browse options for the 2nd file path to allow for new or existing files, and the file open to be for Open or Create.  Without those changes, you can't pick a new file and if you could, you'd get error when it tried to open the file that didn't exist.

  • Copying files to my MBA Early 2011, 10.6.7 changes their 'date modified' to the current date. It happens only to some of the files copied. Any idea what my be causing it. Never had this problem before, on othe macs I have owned.

    Copying files to my new MBA (early 2011) running 10.6.7 results in their 'modification date' to change to the current date. Any idea what may be causing it. It happens only to some of the files (there is no pattern to it.). I have not seen it previously in many other macs I have owned.

    Hi all
    I am bumping this thread again because I can't believe that I am the only one experiencing this issue. At times it is costing me hours repeatedly copying files until eventually the date doesn't show as modified.
    I thought I had cracked it earlier by changing the Sharing & Permissions settings, and applying to enclosed items the ability for Everyone, Me and Staff fo have Read & Write permissions - but that doesn't seemed to have changed much.
    Copying in smaller batches helps sometimes, but then other times it makes no odds, and the Date Modified date shows as the original date for a few seconds, only to change to today's date. Driving me nuts because there is no pattern to it.
    Files which I have copied and show their original modified date, when moved to a subfolder, then have their date changed to today!!! Copying other files into the folder results in files that were previously OK having their Date Modified changed!
    Am I on the only one?

  • Get File by modified date and extension then copy to location

    Hello,
    Ill give a brief description of what I am trying to accomplish. We have a bunch of files that are sitting on a deduplication box (Exagrid), we want to try and copy certain files off of this dedup box, however, if we copy EVERYTHING this will fill up our
    entire NAS Device. So, instead, I need to build a script that finds files by a modified date and extension type, then copy those to a specified folder. 
    Here is the code:
    $logfile = "C:\CopyResults.log"
    function LogWrite{
    Param([string]$logstring)
    Add-Content $logfile -Value $logstring
    $date = get-date
    $DateToCompare = (Get-Date).AddDays(-2)
    $Files = Get-ChildItem -Recurse | Where-Object {$_.LastWriteTime -gt $DateToCompare} | Where-Object {$_.Extension -eq ".xlsx"}
    foreach($File in $Files){
    Write-Host "$File was copied to C:\test"
    Copy-Item $File C:\test
    logWrite "$File was copied to c:\test --- $date"
    The file extension is of course just a test, as well as the destination of where these files are to be copied to. Ultimately I would like to build in the functionality that will check the date and delete the files after a certain amount of time has passed and
    then copy new files. For now, though, I would be satisfied with just some insight on my own idea and know where I am going wrong or how to clean this up a little better. The overall issue I having is when I run this portion of the script I get this: Copy-Item
    : Cannot find path 'C:\Users\<username>\<filename>.xlsx' because it does not exist.
    At line:13 char:5
    +     Copy-Item $File C:\test
    +     ~~~~~~~~~~~~~~~~~~~~~~~
        + CategoryInfo          : ObjectNotFound: (C:\Users\<usern...al filename>.xlsx:String) [Copy-Item], ItemNotFoundException
        + FullyQualifiedErrorId : PathNotFound,Microsoft.PowerShell.Commands.CopyItemCommand

    I solved the issue. Here is the full script:
    net use \\netpath\filename /user:<Username> <Password>
    $logfile = "C:\CopyResults.log"
    function LogWrite{
    Param([string]$logstring)
    Add-Content $logfile -Value $logstring
    $date = get-date
    $DateToCompare = (Get-Date).AddDays(-1)
    $VBMs = Get-ChildItem \\netpath\filename -Recurse | Where-Object {$_.LastWriteTime -gt $DateToCompare} | Where-Object {$_.Extension -eq ".vbm"}
    $VBKs = Get-ChildItem \\netpath\filename -Recurse | Where-Object {$_.LastWriteTime -gt $DateToCompare} | Where-Object {$_.Extension -eq ".vbk"}
    $VIBs = Get-ChildItem \\netpath\filename -Recurse | Where-Object {$_.LastWriteTime -gt $DateToCompare} | Where-Object {$_.Extension -eq ".vib"}
    foreach($VBM in $VBMs){
    Write-Host "$VBM was copied to C:\test"
    Copy-Item $VBM.FullName \\netpath\filename
    logWrite "$VBM was copied to c:\test --- $date"
    foreach($VBK in $VBKs){
    Write-Host "$VBK was copied to C:\test"
    Copy-Item $VBK.FullName \\netpath\filename
    logWrite "$VBK was copied to c:\test --- $date"
    foreach($VIB in $VIBs){
    Write-Host "$VIB was copied to C:\test"
    Copy-Item $VIB.FullName \\netpath\filename
    logWrite "$VIB was copied to c:\test --- $date"
    "In this case you are handing a System.IO.FileSystemInfo.FileInfo object to the Copy-Item cmdlet. I think the cmdlet is probably defaulting to using the .Name property and that is not enough information for the copy to work. When you explicitly give the
    .FullName property to the cmdlet, it now has the information that it needs."
    I found this on Stack Overflow buried in the forums and it resolved all my issues. I have tried using that tool and it works very well, however, Powershell is what I am using for everything and I would like to keep it that way.
    Thank you all for your time and help.

  • Copying file based on initial part of file's name and creation date

    We have a client who dumps data in one of our 2003 SP2 server. We need a script to copy a single file from this location to 
    one of our 2008 R2 server.Like in the source server we have
    abc_Today'sDate_hexadecimal.csv  (File Name)
    Today
                   (Date Modified)
    abc_YesterDay'sdate_hexadecimal.csv      (File Name)
    Tomorrow         (Date Modified)
    'abc' is constant in the file name where rest changes.  We want this file, whose name is starting with 'abc' and created today to be copied to destination, preferably replacing the file which was copied yesterday.
    I know nothing about scripting so any help would be appreciated
    Thanks

    If you are hoping to get a complete solution, you won't as that is not the purpose of this forum.
    If you are hoping to learn how to accomplish this task yourself with a little help, then this is the place for you, but...
    you will be better off to start learning powershell, perhaps using this exercise as a practical example. So to start, you should type these commands in the powershell console:
    help get-childitem
    help copy-item
    if the documentation that this presents makes no sense to you, you will have to get past the "knowning nothing about scripting" condition first. If so, someone here will no doubt provide some usefu references...
    Al Dunbar -- remember to 'mark or propose as answer' or 'vote as helpful' as appropriate.
    Excellent - wel stated. I wil keep this.
    ¯\_(ツ)_/¯

  • How to copy files and last modified date

    Does anyone out there know who to copy files files from one directory to another and retain the date time stamp of the source files?
    I've been able to write some code that copies files from one directory to another but I always end up with a Date Last Modified that is based on the current date. I want to retain the date time stamp from the original file that I copied. Is this possible in Java?
    Mike

    Call file.setLastModified() after you close it.

  • Hyperion DIM (Data integn mgmt/Informatica) - Create Workflow to copy files

    Hello,
    I understand this is an ODI forum but I could not find a Oracle Hyperion DIM forum and so I thought some of you might know DIM. This should be easy for those who have Informatica experience.
    So we are using WinSCP for copying files from one folder to the other.
    Those who do not know what WINSCP is - it is a simple to tool used to conduct secure file transfer b/w local and remote desktop.
    We are doing this file copy from one folder to the other on a repeat basis and we would like to automate the workflow using DIM.
    Well, the task is that I basically log on to WinSCP by providing a hostname, username and password and from a specific folder I have to pick up a file and drop it into a different folder.
    I need to create a workflow to do this so that it may be automated. Can anybody suggest ideas on which tasks I may use within DIM and the general approach? This would be great help. Thanks.
    -- Adi

    Hello,
    I understand this is an ODI forum but I could not find a Oracle Hyperion DIM forum and so I thought some of you might know DIM. This should be easy for those who have Informatica experience.
    So we are using WinSCP for copying files from one folder to the other.
    Those who do not know what WINSCP is - it is a simple to tool used to conduct secure file transfer b/w local and remote desktop.
    We are doing this file copy from one folder to the other on a repeat basis and we would like to automate the workflow using DIM.
    Well, the task is that I basically log on to WinSCP by providing a hostname, username and password and from a specific folder I have to pick up a file and drop it into a different folder.
    I need to create a workflow to do this so that it may be automated. Can anybody suggest ideas on which tasks I may use within DIM and the general approach? This would be great help. Thanks.
    -- Adi

  • Performing a copy of a "live" data file?

    Hello -
    Is there a risk (to the source data file) of performing a copy command on a live data file? My sys admin informed me he was copying a data file. He was not concerned about the file on the destination drive being corrupted. He just had to fill up another file system to perform a test. My concern was any potential corruption to the "live" DB file.
    Thanks,
    mike

    I agree that copying a file should be 0% risk of corruption on the source datafile.
    Let's think about this. How does a conventional backup work? We put the tablespace in backup mode, and then use O/S level tools to make the copy. All that putting the tablespace in backup mode does is write to the datafile header of each datafile in the tablespace, and increases I/O by writing some full blocks to the redo log. After that, the O/S level copy simply copies the datafiles. If someone comes along and just copies the datafiles, the O/S copy mechanism is the very same as if you were doing a conventional backup. There is no difference.
    So, if you believe that copying a file can damage the source file, then how do you reliably backup your database files without having to resort to a cold backup?
    Copying files doesn't damage the source file. Having said that, I agree that the sysadmin should not be messing with Oracle datafiles. Today he uses the files as source to copy from, tomorrow he removes, renames, whatever. Teach him to use dd to create large files.
    dd if=/dev/zero of=/tmp/my_2gb_file bs=1048576 count=2048
    Hope that helps,
    -Mark

Maybe you are looking for