Importing csv with dates

Hi all,
I am importing csv files that contain (among others) two columns with dates into Numbers. The dates are all uniformly formatted by dd.mm.yyyy, which is in Germany the usual date format. However, the import changes them into numbers, e.g.
26.10.2009 becomes 38650
13.11.2009 becomes 38668
So, there seams to be some logic behind, but how can I conveniently get the correct date back?
Thanks for any help.
Best regards.

Unfortunately, I wasn't able to change the cell format type to 'date' *after* importing the data, it just remained 'number'.
However, I found a way to work around the problem: Introducing a new column and using there the function EDATE gave me the original date back! The function takes two arguments, the first I let point to the cell with the strange number, the second I let point to an empty cell.
EDATE might not be meant to work that way and this feature is not documented, anyway, it worked for me.

Similar Messages

  • Import csv with date into HANA

    Hi,
    I have a table with multiple text columns and one date column in the csv I am trying to import I have the date as YYYY-MM-DD which I thought was the format I needed. Every time I select the csv and the table to import to then press next my whole studio freezes up and I have to do a force close.
    I attached the screen where I get to when it freezes. When I take this column out it loads fine.
    Can someone tell me the correct format for date to enter it in CSV, or what I am doing wrong?
    Thanks,
    Connor

    You may want to open a support incident for this.
    - Lars

  • Import CSV, Edit Data, Export CSV

    Hello,
    I am a powershell novice and have been given a task to complete. I need to scan through AD, find all of the servers that are supposed to be active, "ping" them and, if active, get their IP address and verify that a number of services are running.
    Once all of the data is collected, I need to write all of this out to a .CSV file. So far, what I have is this:
    Using Import-Module ActiveDirectory, I am able to get the full list of servers with appropriate information and export that to a CSV
    I import the CSV file above back in and create new headers (columns) for more information
    The full list of servers are pinged individually. If pingable, I get the IP address and service information needed
    Where I am stuck is writing all of this back out to a .CSV file. Right now, I am just writing the information back out to the screen for verification. I'm sure it's something small, but being a novice I'm not sure where to go. Any help?

    Here is the full code. The Write-Host statements are just for me during the testing to see if the information is correct. The last statement was what I tried to write the information out, but it came up empty and I really haven't found a way to get it to
    work.
    Import-Module ActiveDirectory    
    Get-ADComputer -LDAPFilter "(OperatingSystem=*Server*)" -Property * |
        Select-Object Name,Created,LastLogonDate,OperatingSystem,OperatingSystemServicePack,OperatingSystemVersion,CanonicalName |
        ? {$_.CanonicalName -notlike "*Archived Servers*" -and $_.CanonicalName -notlike "*Sun Prod*"} |
        Export-CSV "c:\temp\AllADComputers4.CSV" -NoTypeInformation -Encoding UTF8
    #Add other headers
    $csv = get-content "c:\temp\AllADComputers4.CSV"
    $csv[0] += ",Pingable,IPAddress,SCCM Status,AV Name,AV Version,AV Status"
    $csv | out-file "c:\temp\AllADComputers4.CSV"
    $MachineInfo = Import-Csv -Path "c:\temp\AllADComputers4.CSV"
    $Processed = ForEach ($objitem in $MachineInfo)
        If (Test-Connection -ComputerName $Objitem.Name -Quiet -Count 1)
            If (Get-Service -name "Symantec AntiVirus" -computername $Objitem.Name)
                $AVStatus = Get-Service -name "Symantec AntiVirus" -computername $Objitem.Name
                $AVName = "Symantec AntiVirus"
            Elseif (Get-Service -name SepMasterService -computername $Objitem.Name)
                $AVStatus = Get-Service -name SepMasterService -computername $Objitem.Name
                $AVName = SepMasterService
            ElseIf (Get-Service -name "Norton AntiVirus Server" -computername $Objitem.Name)
                $AVStatus = Get-Service -name "Norton AntiVirus Server"-computername $Objitem.Name
                $AVName = "Norton AntiVirus Server"
            Else
                $AVStatus = "Stopped"
                $AVName = "None"
            $SCCMAgentStatus = Get-Service -name ccmexec -computername $Objitem.Name
            $Networks = Get-WmiObject Win32_NetworkAdapterConfiguration -ComputerName $Objitem.Name | ? {$_.IPEnabled}
            $objitem.Pingable = "Yes"
            foreach ($Network in $Networks)
                    $IPAddress  = $Network.IpAddress[0]
            Write-Host $Objitem.Name,"...Success...",$IPAddress,"   SCCM Status = ",$SCCMAgentStatus.Status, "Anti-Virus = ",$AVName, "Status = ",$AVStatus.Status
            $service = gwmi -computername $objItem.Name -class Win32_Service | ? {$_.Name -eq $AVName}
            $path = ($service | Select -Expand PathName)
            $path = $path.Replace(":","$")
            $path = $path.Replace('"','')
            $FullPath = "\\"+$objitem.Name+"\"+$path
            $ver = Get-ItemProperty -Path $FullPath | select-object -expand VersionInfo | select-object -expand ProductVersion
            Write-Host $Fullpath
            Write-Host "Version = ",$ver
        Else
            Write-Host $Objitem.Name"...Failed"
            $objitem.Pingable = "No"
    $Processed | export-csv -Path "c:\temp\AllADProcessed.CSV"

  • Numbers issue importing csv with this field "00:00:00:00"

    Greetings,
    I'm trying to import a .csv file into Numbers. One of the fields is Timecode with the format 00:00:00:00
    Numbers tries to read this as a date and doesn't know how to deal with it. I've tried every cell type to no avail..
    Anyone have any ideas?
    thx

    I'm importing a .csv file from an audio recorder with the goal of generating a .pdf report. Inside the .csv file is a column that for every file has a smpte timecode field in the form of hr:min:sec:frames or 01:02:45:00
    When importing this into Numbers, the program can't convert this properly so it creates a useles date out of if. I need Numbers to just import this as Text or not attempt to convert it to some other field type. Here's what a row from the output of the recorder looks like
    25Y-04.WAV,25Y,04,00:01:35,04:44:12:00,"MIX",,"BOOM-416",,,,,,,,,,"",
    Numbers doesn't deal with 04:44:12:00

  • Import CSV/Excel data into Flex Datagrid using Java/JSP

    Hi,
    Is there any way we can import excel/csv data into flex datagrid component.
    I see that is been done using php in this site. http://code.google.com/p/as3xls/
    Any help would be appreciated.
    Thanks in advance

    You can store the list of data from the Excel sheet that you have read already into an ArrayList
    Then go through the JDBC tutorial, and learn how to insert a single record into the table.
    After you understand how to insert a single record, you can iterate through the ArrayList of data and insert them into the table in the database.
    If you know JDBC, see if you can use OR Mapping frameworks like Hibernate or iBatis, they make it very easy to manage database in Java.

  • Importing Events with dates for titles instead of "untitled event"?

    Hi everyone,
    My wife has a simple wish - she'd like her photos organized into folders (in the case of iPhoto, events) titled by the date those photos were taken.  So for example
    2011_07_01
    Or some similar variant.
    Now, we've given up trying to get that to work with importing in Aperture - it's just bizarrely difficult to figure out how to do something so simple, I can't fathom it.
    So we went back to iPhoto. The wierd thing is, this worked for a large directory of photos we imported from our old computer. But when we try importing new photos from her camera - a Canon Powershot 300 HS - it just creates a whole bunch of Untitled Events. The weird thing is, they are broken up properly into separate days, but it doesnt name the event by the date which makes no sense. My hope is a person doesn't have to manually rename every single event as the date - I kinda figure we use this sort of software to avoid that kind of thing.
    Any tips or workarounds you guys know of?
    Best Regards
    Jarrett

    What you are are describing is the way iPhoto works - in the iPhoto preferences you can set the parameters for spliting imports into events but you will ahve to add the event titles yourself - titling an event with the photos dates is no an option
    iPhoto menu ==> provide iPhoto feedback to suggest to Apple
    And folders in iPhoto do not hold photos - they hold albums and other folders
    LN

  • How to import xml with data?

    Excel 2013
    i got an XML file generated by the GPinventory tool. when i import it into Excel, i only get the headers.
    how can i import the data as well?

    Hi,
    As far as I know, the GPinventory tool Supported Operating System :Windows 2000, Windows Server 2003, Windows XP, I suppose it generated XML file with Excel 2003 schema, but Excel 2013 have a new XML schema. This issue usually caused by the XML scheme.
    Please create a Excel 2013 schema XML file
    to test with Excel 2013, if the file could be import well. We'd better check the generated XML file.
    As a workaround, we'd better save the results as
    a tab-delimited text file and then import to Excel 2013.
    If this issue still exists, please upload the XML file through OneDrive. I want to test it.
    Regards,
    George Zhao
    Forum Support
    Come back and mark the replies as answers if they help and unmark them if they provide no help. If you have any feedback on our support, please click "[email protected]"

  • W520 - Lenovo Password Manager - Can .CSV or .DAT files be imported from other Password Managers?

    The title says it all!
    Currently using Norton Identity Safe's standalone password manager, but would like to start using Lenovo's instead. Attempted to import .DAT and/or .CSV files from Norton, but no option for these file types appeared in Lenovo's Password Manager.
    If not an option to import directly, is it possible to add one of the files to the location Lenovo stores the password manger entries?
    Solved!
    Go to Solution.

    Hello, 
    No you can't import .CSV or .DAT files into Password Manager. It uses a .PWM file format, and uses a 256-bit key Advanced
    Encryption Standard (AES) with Microsoft CryptoAPI. You can export your info to be used in another machine with the Thinkvantage Password Manager installed, but not from another password manager.
    More info here:
    Password Manager 4 Deployment Guide
    Cheers!
    ThinkPad W540 (20BG) - i7-4800MQ/24GB // ThinkPad T440s (20AQ) - i7-4600U/12GB
    ThinkPad T440p (20AW) - i7-4800MQ/16GB // ThinkPad Helix (3698-6EU) - i5-3337U/4GB
    ThinkPad W520 (4282-W4Q) - i7-2720QM/32GB // ThinkPad T400 (2767-W1C) - P9500/8GB
    ThinkPad T61 (7665-CTO) - T7700/4GB // ThinkPad T60p (8741-C2G) - T7400/4GB

  • PowerShell Import-CSV cmdlet Help

    "as I am assuming I need the field to be EmailAddress"That is absolutely correct, to pipe the object from a csv straight to new-aduser the column headers in the csv must match the parameters of new-aduser. So mail won't work, but EmailAddress will.You can do the following to get the parameters that are available.Powershellget-help new-aduserIf changing the csv column headers isn't going to work then you can build the script like this:PowershellForEach($user in (Import-CSV .\data.csv) ) \{ New-ADUser -EmailAddress $user.mail ......}

    PowershellForEach($user in (import-csv .\FreshmanADImport2.csv) ){ $userobj = $user | New-ADUser -PassThru -Path "OU=Freshman,OU=Students,DC=domain,DC=com" "Students","Student_iPads" | Add-ADGroupMember -Members $userobj}I haven't looked at your csv yet, but here is the gist of what you are after.Notes:1. Loop thru them with a ForEach() will give you the control you need to do the groups.2. -PassThru lets you capture the new user3. -Path lets you specify where the user is created.

  • Export with data length semantics

    Hello,
    I've following problem.
    I have a table abcd which contains 2 VARCHAR2 columns with different data length semantics (one with BYTE, one with CHAR). Charset is Single Byte; let's say WE8MSWIN1252, so data length semantics should not be a problem. should not. details later.
    So this would be:
    create table abcd (a_char VARCHAR2(2 CHAR), a_byte VARCHAR2(2 BYTE));after that I export the table via exp. I'm not setting NLS_LENGTH_SEMANTICS environment variable, so BYTE is used.
    In the dump file the data length semantics for the byte col is omitted, as I exported it with BYTE:
    create table abcd (a_char VARCHAR2(2 CHAR), a_byte VARCHAR2(2));after that, I "accidently" import it with data length semantics set to CHAR, and the table looks like this now
    abcd
    a_char VARCHAR2(2 CHAR)
    a_byte VARCHAR2(2 CHAR)Same happens vice versa when using CHAR for export and BYTE for import...
    In single byte charsets this might not be so much of a problem, as one CHAR is equal to one BYTE, but...
    If I compile plsql against the original table, and run against the outcoming table after export, I get an ORA-4062, and I have to recompile...
    Would not be a problem if the plsql I compile would be on the database...Big problem is that the ORA-4062 occurs in forms, where it's difficult for me to recompile (I would have to transfer all the sources to customer and compile there).
    Is there any possibility to export data length semantics regardless which environment variable is set?
    database version would be 9.2.0.6; but if there exists a solution in higher versions I would also be happy to hear them...
    many thanks,
    regards

    I can't reproduce your problem:
    SQL> show parameter nls_length_semantics
    NAME                                 TYPE        VALUE
    nls_length_semantics                 string      BYTE
    SQL> create table scott.demo( col1 varchar2(10 byte), col2 varchar2(10 char) );
    SQL> describe scott.demo
    Name                                      Null?    Type
    COL1                                               VARCHAR2(10)
    COL2                                               VARCHAR2(10 CHAR)
    $ export NLS_LENGTH_SEMANTICS=BYTE
    $ exp scott/tiger file=scott.dmp tables=demo
    SQL> drop table scott.demo;
    $ export NLS_LENGTH_SEMANTICS=CHAR
    $ imp scott/tiger file=scott.dmp
    SQL> describe scott.demo
    Name                                      Null?    Type
    COL1                                               VARCHAR2(10 BYTE)
    COL2                                               VARCHAR2(10)
    SQL> alter session set nls_length_semantics=byte;
    SQL> describe scott.demo
    Name                                      Null?    Type
    COL1                                               VARCHAR2(10)
    COL2                                               VARCHAR2(10 CHAR)Can you post a test like mine?
    Enrique
    PS If you have access to Metalink, read Note:144808.1 Examples and limits of BYTE and CHAR semantics usage. From 9i and up, imp doesn't read nls_length_semantics from the environment.
    Edited by: Enrique Orbegozo on Dec 16, 2008 12:50 PM
    Edited by: Enrique Orbegozo on Dec 16, 2008 12:53 PM

  • Excel issues with importing CSV or HTML table data from URL - Sharepoint? Office365?

    Greetings,
    We have a client who is having issues importing CSV or HTML table data as one would do using Excel's Web Query import from a reporting application.  As the error message provided by Excel is unhelpful I'm reaching out to anyone who can help us begin to
    troubleshoot problems affecting what is normal standard Excel functionality.  I'd attach the error screenshot, but I can't because my account is not verified....needless to say it says "Microsoft Excel cannot access  the file https://www.avantalytics.com/reporting_handler?func=wquery&format=csv&logid=XXXX&key=MD5
    Where XXXX is a number and MD5 is an md5 code.  The symptoms stated in the error message are:
    - the file name or path does not exist
    -The file is being used by another program
    -The workbook you are trying to save has the same name as a currently open workbook.
    None of these symptoms are the case, naturally. The user encountered this with Excel2010, she was then upgraded to Excel2013 and is still experiencing the same issue. The output of this URL in a browser (IE, Chrome, Firefox) is CSV data for the affected
    user, so it is not a network connectivity issue.  In our testing environment using both Excel2010 or 2013 this file is imported successfully, so we cannot replicate.  The main difference I can determine between our test environment and the end-user
    is they have a Sharepoint installation and appear to have Office365 as well.
    So,  my question might more appropriately be for Sharepoint or Office365 folks, but I can't be sure they're  a culprit.  Given this - does anyone have any knowledge of issues which might cause this with Sharepoint or Office365 integrated with
    Excel and/or have suggestions for getting more information from Excel or Windows other than this error message?  I've added the domain name as a trusted publisher in IE as I thought that might be the issue, but that hasn't solved anything.  As you
    can see its already https and there is no authentication or login - the md5 key is the authentication.  The certificate for the application endpoint is valid and registered via GoDaddy CA.
    I'm at a loss and would love some suggestions on things to check/try.
    Thanks  -Ross

    Hi Ross,
    >> In our testing environment using both Excel 2010 and 2013 this file is imported successfully, so we cannot replicate.
    I suspect it is caused by the difference of web server security settings.
    KB: Error message when you use Web query to a secure Web page (HTTPS://) in Excel: "Unable to open"
    Hope it will help.
    By the way, this forum is mainly for discussing questions about Office Development (VSTO, VBA and Apps for Office .etc.). For Office products feature specific questions, you could consider posting them on
    Office IT Pro forum or Microsoft Office Community.
    Regards,
    Jeffrey
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Importing CSV file with Data Merge Fails

    Specs
    See pasted text from CSV at http://pastebin.com/mymhugpN
    I am using InDesign CS6 (8.0.1)
    I created the CSV by downloading it from a Google Spreadsheet as a CSV. I confirm with the Terminal that the character encoding is utf-8 usnig the file command.
    Problem detailed
    I am trying to import a CSV file (utf-8) with Data Merge via the Select Data Source... command with Show Import Options checked. When viewing the Data Source Import Options dialog, I set the following options—Delimiter:Comma, Encoding:Unicode, Platform:Macintosh. I leave Preserve Spaces in Data Source unchecked. It fails to import any variables and produces no error message. I have tried other CSV files as well (created TextEdit, Espresso, etc.) and it seems that InDesign will not import any files if Unicode is specified as the encoding, no matter which other options are specified.
    Can anyone else confirm this?
    Importing as ACSII works, but obviously does not display my content correctly.

    Mike is having some trouble posting in this thread (and I am too), but he sent me a PM with what he wanted to say:
    OK. I think I might have a positive answer for you.
    I was getting lost in the upper ASCII characters you showed. In your test file I never could see any--a case of not seeing the trees for the forest.
    Your quote marks are getting dropped in your test file. Now, this may or may not affect other factors but it does in some further testing. I believe ID has an issue with dropping quote marks even in a plain ASCII file if the marks are at the beginning of a sentence and the file is tab delimited. Call it a bug.
    Because of all the commas and quote marks in your simple file, I think you should be exporting from Google Docs' spreadsheet as a tab-delimited file. This exported file has to be opened in a text editor capable of saving it out as a UTF-16 BE (Big Endian) type of file.
    Also, I think you are going to have to use proper quote marks throughout, or change them in the exported tab-delimited file. Best to have a correct source, though.
    Here is your sample ZIPped up. I think it works properly. But then again, I think I might be bleary-eyed by now.
    http://www.wenzloffandsons.com/temp/merge_psalms_utf-16.zip
    Take care, Mike

  • Using PowerShell to import CSV data from Vendor database to manipulate Active Directory Users

    Hello,
    I have a big project I am trying to automate.  I am working in a K-12 public education IT Dept. and have been tasked with importing data that has been exported from a vendor database via .csv file into Active Directory to manage student accounts. 
    My client wants to use this data to make bulk changes  to student user accounts in AD such as moving accounts from one OU to another, modifying account attributes based on State ID, lunchroom ID, School, Grade, etc. and adding new accounts / disabling
    accounts for students no longer enrolled.
    The .csv that is exported doesn't have headers that match up with what is needed for importing in AD, so those have to be modified in this process, or set as variables to get the correct info into the correct attributes in AD or else this whole project is
    a bust.  He is tired of manually manipulating the .csv data and trying to get it onto AD with few or no errors, hence the reason it has been passed off to me.
    Since this information changes practically daily, I need a way to automate user management by accomplishing the following on a scheduled basis.
    Process must:
    Check to see if Student Number already exists
    If yes, then modify account
    Update {School Name}, {Site Code}, {School Number}, {Grade Level} (Variables)
    Add correct group memberships (School / Grade Specific)
    Move account to correct OU (OU={Grade},OU=Students,OU=Users,OU={SiteCode},DC=Domain,DC=net)
    Remove incorrect group memberships (School / Grade Specific)
    Set account status (enabled / disabled)
    If no, create account
    Import Student #
    Import CNP #
    Import Student name
    Extract First and Middle initial
    If duplicate name exists
    Create log entry for review
    Import School, School Number, Grade Level
    Add to correct Group memberships (School / Grade Specific)
    Set correct OU (OU={Grade},OU=Students,OU=Users,OU={SiteCode},DC=Domain,DC=net)
    Set account Status
    I am not familiar with Powershell, but have researched enough to know that it will be the best option for this project.  I have seen some partial solutions in VB, but I am more of an infrastructure person instead of scripting / software development. 
    I have just started creating a script and already have hit a snag.  Maybe one of you could help.
    #Connect to Active Directory
    Import-Module ActiveDirectory
    # Import iNOW user information
    $Users = import-csv C:\ADUpdate\INOW_export.csv
    #Check to see if the account already exists in AD
    ForEach ( $user in $users )
    #Assign the content to variables
    $Attr_employeeID = $users."Student Number"
    $Attr_givenName = $users."First Name"
    $Attr_middleName = $users."Middle Name"
    $Attr_sn = $users."Last Name"
    $Attr_postaldeliveryOfficeName = $users.School
    $Attr_company = $users."School Number"
    $Attr_department = $users."Grade Level"
    $Attr_cn = $Attr_givenName.Substring(0,1) + $Attr_middleName.Substring(0,1) + $Attr_sn
    IF (Get-ADUser $Attr_cn)
    {Write-Host $Attr_cn already exists in Active Directory

    Thank you for helping me with that before it became an issue later on, however, even when modified to be $Attr_sAMAaccountName i still get errors.
    #Connect to Active Directory
    Import-Module ActiveDirectory
    # Import iNOW user information
    $Users = import-csv D:\ADUpdate\Data\INOW_export.csv
    #Check to see if the account already exists in AD
    ForEach ( $user in $users )
    #Assign the content to variables
    $Attr_employeeID = $users."Student Number"
    $Attr_givenName = $users."First Name"
    $Attr_middleName = $users."Middle Name"
    $Attr_sn = $users."Last Name"
    $Attr_postaldeliveryOfficeName = $users.School
    $Attr_company = $users."School Number"
    $Attr_department = $users."Grade Level"
    $Attr_sAMAccountName = $Attr_givenName.Substring(0,1) + $Attr_middleName.Substring(0,1) + $Attr_sn
    IF (Get-ADUser $Attr_sAMAccountName)
    {Write-Host $Attr_sAMAccountName already exists in Active Directory
    PS C:\Windows\system32> D:\ADUpdate\Scripts\INOW-AD.ps1
    Get-ADUser : Cannot convert 'System.Object[]' to the type 'Microsoft.ActiveDirectory.Management.ADUser'
    required by parameter 'Identity'. Specified method is not supported.
    At D:\ADUpdate\Scripts\INOW-AD.ps1:28 char:28
    + IF (Get-ADUser $Attr_sAMAccountName)
    + ~~~~~~~~~~~~~~~~~~~~
    + CategoryInfo : InvalidArgument: (:) [Get-ADUser], ParameterBindingException
    + FullyQualifiedErrorId : CannotConvertArgument,Microsoft.ActiveDirectory.Management.Commands.GetAD
    User

  • How to import csv file with multiple tables into sql server

    I have multiple csv files that has one sheet but has 130 headers with each header having different data. 
    I'd like to import each one of these header rows with data into its own file in sql server. 
    I know very basic SSIS and am but am not familiar with the scripting in it though which what I assume I'd have to use. 
    Each header in the csv file is structured as such(also see example pic):
    first header would be this:                             
          ITEM = ORG_V                              
          DATE = 2013-07-22 10:00 ~ 2013-07-22 10:15      
    column names
    data
    second header would be this:
    ITEM = TER_V
          DATE = 2013-07-22 10:00 ~ 2013-07-22 10:15
    column names
    data
    The headers can be at any random row number as well as the data size in each excel file differs but they all start with "ITEM ="
    and then in the next row "DATE ="
    I could also convert these to excel files if it makes this process easier. 

    Why don't you put a filter on D3, filter out the blanks, copy/paste to a new CSV file, save it, and import it.
    There's no way you're going to get SQL to do that kind of thing for you.  The language is for set based operations, not for complex data manipulation tasks.
    Knowledge is the only thing that I can give you, and still retain, and we are both better off for it.

  • Wrong date format when import CSV files

    When you import a CSV file that contains fields with German date formats, these fields are displayed incorrectly.
    Example: Contents of the CSV file "01.01.14". After importing the corresponding cell in Numbers has the content "40178".
    A reformat the cell to a date format is not possible.
    How do I get the date in the new version of Numbers displayed correctly?

    It seems that there are more than a few problems related to import/export with non-US localizations.
    I, personally, don't have a solution to your problem. I started to adjust my Language & Region settings to test your problem but it was several settings, I didn't get it right, and I didn't want to mess up my computer so I set everything back to US/English.
    The only workarounds I can suggest are
    Insert a new column into your table and in it put a formula that adds the number to the date 01.01.1904.  Or,
    Edit the CSV in TextEdit to Replace All "." with "/".  This will work if "." is used for nothing else but these dates.
    I recommend the second one if it will work for you. Hopefully Apple is addressing problems such as the one you are seeing.

Maybe you are looking for