Wavefrom graph time from .csv

Hello everyone
I use "read from spreadsheet file" to read a two column .csv file in a waveform graph.
Time is defined in the first column and the Data is defined in the second column.
The problem occurs in the x-axis of the graph (Time-axis). It doesn't match with the numbers from the .csv as shown on the picture below, but the Y-axis has correct values.
The VI is uploadet
Best regards
Oesen
Solved!
Go to Solution.
Attachments:
2-Column CSV.vi ‏13 KB

Works fine here once I configure it to use a comma as decimal delimiter (using format=" %,;%.3f " on read from spreadsheet file).
You seem to use a foreign version so that should not be necessary, depending on the language setting of your computer but it seems to be the problem.
As a first step, you should display the entire 2D array in an array indicator to ensure that your number don't get truncated somehow. If they get truncated, dt will be set as zero and thus ignored and taken as 1.
LabVIEW Champion . Do more with less code and in less time .
Attachments:
SetDt.png ‏23 KB

Similar Messages

  • Wavefrom graph time from .csv file

    Hello everyone. I had this problem something about three weeks ago:
    http://forums.ni.com/t5/LabVIEW/Wavefrom-graph-time-from-csv/td-p/2256754
    and solved it by changing the delimiter to a comma decimal delimitter with the format %,;%.2f 
    But this problem occurs again suddenly!.
    I have attached the VI and the .csv file as a ZIP format
    Best regards
    Oesen
    Attachments:
    Trykmaaling_READ.vi ‏36 KB
    ekstra.zip ‏1 KB

    I have deleted "row 0" and the graph is better now, but is still incorrect
    Best regards
    Oesen

  • Loading 361000 records at a time from csv file

    Hi,
    One of my collegue loaded 361000 records from one file file , how is this possible as excel accepts 65536 records in one file
    and even in the infopackage the following are selected what does this mean
    Data Separator   ;
    Escape Sign      "
    Separator for Thousands   .
    Character Used for Decimal Point   ,
    Pls let me know

    hi Maya,
    it just possible, other than ms-excel, we have editor like Textpad that support more 65k rows (and windows Notepad), the file may be generated by program or edited outside in excel, or newer version of excel is used, ms-excel 2007 support more 1 million rows.
    e.g we have csv file
    customer;product;quantity;revenue
    a;x;"1.250,25";200
    b;y;"5.5";300
    data separator ;
    - char/delimeter used to separate field, e.g
    escape sign, e.g
    - "1.250,25";200 then quantity = 1.250,25
    separator for thousands = .
    - 1.250,25 means one thousand two hundred ...
    char used for decimal point
    - - 1.250<b>,</b>25
    check
    http://help.sap.com/saphelp_nw70/helpdata/en/80/1a6581e07211d2acb80000e829fbfe/frameset.htm
    http://help.sap.com/saphelp_nw70/helpdata/en/c2/678e3bee3c9979e10000000a11402f/frameset.htm
    hope this helps.

  • Read from csv file and graph at certain interval

    Hi,
    I've been trying all day to modify a labView program so it reads in data from my csv file to graph temperatures from 4 ovens. The temperatures are logged in the csv file at 3s intervals and i wish to display them on 4 graphs. I've written a vi (or modified one i should say) to take in a temperature value from a numeric control. Im just having extreme difficulty with the reading part.
     If someone can help me it would be greatly appreciated.
    Best regards,
    Dawn
    Attachments:
    Oven temps.csv ‏2 KB
    Oven_level_meas.vi ‏58 KB

    Hi there,
    I have a similar problem here.  I see your point that you should read the file at once but I don't seem to be able to manage it.  Shouldn't you use some kind of
    a buffer to achieve this?  Let me tell you what I am trying to do:  I have a csv file with several columns that I want to display in several waveform charts.  Each column is a wavechart.  I should be able to change the speed at which the data is displayed.  I am able to do this because I change the wait until (ms) in the for loop.  However the application is too slow.  I think because I open the csv file in the for loop and display it while I am reading it.  I have tried to put it outside the for loop but this doesn't work.  I think I need some kind of a buffer in which the data is stored after reading it.  And then this buffer should be used in the for loop.  I have looked at the producer-consumer structure but I dont fully understand it as I am not that experienced with labview.  Any help or changes to my VI would be appreciated.  I have attached the VI and the file with meaningless data I am using for testing.
    Thanks
    Attachments:
    read_data_file_dynamically.vi ‏39 KB
    test.CSV ‏332 KB

  • How to convert time from (hh:mm:ss) to (hhmmss) in CSV file

    Hi Experts,
    I have a CSV file where there is time column which displays time like hh:mm:ss
    But the input I should give for loading the data is hhmmss
    The amount of records that I have is more than 2 lakhs. I tried many ways to convert the time but I'm unsuccessful.
    Can anyone tell me how do I convert the format please.
    Thanks

    Hi,
    You can create a custom infoobject length 6 (say mytime) to read the data from csv file.
    In the field routine assign the 6 character length info object to the required time IO(system time) with routine.
    Inside the field routine...
    result = source_fields-mytime.
    The system will take care of conversion as per the user master setting.
    Thanks,
    Jugal.

  • Controling wavefrom graphs and instrument display.

    Hello, 
    I'm tring to control an old Optical Spectrum Analyzer (OSA) using VISA.
    I manage to read and write and present the date - but not to see the OSA display screen!  I will explain:
    The OSA sweep over some spectrum range continuously - the display actually shows a video. 
    when the sweep ends it store the data in an array - which I read and show on a wavefrom graph.  This is where the OSA display and my waveform link breaks.
    for example if I input a command to change the scale of the y-axis or to insert a marker, the OSA display is changed but not my waveform.  I could not find any other way to read the data.
    from your exprience - such instruments (mine is 20 years old) offer a way to read the entire screen data somehow?
    if not, how can I control my waveform graph through my program - for example define the y scale to be the same as that of the instrument?

    MY OSA is HP70951A.
    I have programming manual of similar item for which most commands work - no driver is available, I've search.
    I have a cluster of x0, dx and Y and I read the data correctly -  so far so good, but I do not see the instrument display at is.
    for example: I could place a marker on my instrument display - but I wouldn't see it in my computer since I read only the y-axis data.
    you can say - so read the marker as well, but what is the point - In theory I can seat in front of my computer screen and add marker on my waveform alone without the instrument. same as changing the y scale division - it will be changed on the instrument display or computer display seperatly.
    I don't know if I'm asking too much....or this is how it should be since this is the first time I write a driver with labview.
    P.S
    I also ask about program control on waveform display in another forum - it is not so trivial for example to change the scaling based on input values in the program comapre to do it with some mouse click.  there are many properites node which are not well explained.

  • Loading data from .csv file into existing table

    Hi,
    I have taken a look at several threads which talk about loading data from .csv file into existing /new table. Also checked out Vikas's application regarding the same. I am trying to explain my requirement with an example.
    I have a .csv file and I want the data to be loaded into an existing table. The timesheet table columns are -
    timesheet_entry_id,time_worked,timesheet_date,project_key .
    The csv columns are :
    project,utilization,project_key,timesheet_category,employee,timesheet_date , hours_worked etc.
    What I needed to know is that before the csv data is loaded into the timesheet table is there any way of validating the project key ( which is the primary key of the projects table) with the projects table . I need to perform similar validations with other columns like customer_id from customers table. Basically the loading should be done after validating if the data exists in the parent table. Has anyone done this kind of loading through the APEX utility-data load.Or is there another method of accomplishing the same.
    Does Vikas's application do what the utility does ( i am assuming that the code being from 2005 the utility was not incorporated in APEX at that time). Any helpful advise is greatly appreciated.
    Thanks,
    Anjali

    Hi Anjali,
    Take a look at these threads which might outline different ways to do it -
    File Browse, File Upload
    Loading CSV file using external table
    Loading a CSV file into a table
    you can create hidden items in the page to validate previous records before insert data.
    Hope this helps,
    M Tajuddin
    http://tajuddin.whitepagesbd.com

  • Importing users into WGM from csv file issues/crash

    Hi,
    i've been importing user information from csv files into WGM via the +server >  import+ function .
    It worked the first few times but now when i try the import progress bar pops up and promptly disappears without any thing importing.
    i've tried restarts, new admin account, reinstalled  WGM.
    I've also trashed some pref but i don't really know which ones i should be losing.
    The servers an OD master.
    any help would be appreciated.
    as a last resort what do i need to backup/save if i were to format/reinstall osx server? keeping my settings etc.....
    thanks
    paul

    What I did was:
    Exported the user list, to create an XML file in the correct format.
    Using this format, I created a spreadsheet in Excel (sorry Apple), and in the final column I created a field that concatenated the information I wanted in the ':' deliminated format of the previously export XML.
    Then just copy and past via pico into a pure text file and imported that.
    You have to be careful with comments in Passenger, using special characters (';!@#$%^ and others can cause the WGM to fail and crash.

  • Initial load of inventory level from csv - double datarows in query

    Hello everybody,
    a query result shown in a web browser seems strange to me and I would be very glad, if anyone can give me some advice how to solve the problem. As I do not think that it is related to the query, I posted it into this forum.
    The query refers to an InfoCube for inventory management with a single non-cumulative key figure and two other cumulative key figures for increase and decrease of inventory. The time reference characteristic is 0CALDAY. The initial load has been processed reading from a flat file (CSV), the structure looks like this:
    Product group     XXX
    Day               20040101
    Quantity          1000
    Increase          0
    Decrease          0
    Unit               ST
    The initial load runs fine, the system fills all the record sets into the InfoCube. Unfortunately I do not know how to look at the records written into the cube, because only the cumulative key figures are shown in InfoCube-> Manage-> Contents.
    Well, when executing the query, a really simple one, the result is just strange, because somehow there are now two rows for each product group with different dates, one with the 1st of January, 2004 and the other for the 31st of December, 2003 containing both 1000 units. The sum is 2000.
    It became more confusing, when I loaded the data for increase and decrease: now the quantities and sums      are correct, but the date of the initial load is a few days later than before, the data table in the query does not contain the 1st of January.
    Does anybody know, what I did wrong or where there is information about how to perform an initial load of inventory from csv in a better way?
    Kind regards
    Peter

    Peter,
    Inventory is not that straight forward to evaluate as it is non-cumulative. Basically it means that one KF is derived from one/two other KFs. You cannot see non-cumulative KFs in manage infocube.
    Have you uploaded opening balances separately? If so, your data for 31st of december is explained.
    In non-cumulative cubes, there need not be a posting for a particular day for a record to exist. For e.g. if you have stock as 10 units on 1st and then no posting for 2nd and 3rd and then increase 10 units on 4th, even for 2nd and 3rd, the non-cumulative KF will report as 10 units (stock on 1st rolled forward).
    There is a how to...inventory management document on service market place that explains this quite nicely.
    Cheers
    Aneesh

  • Import Meta Data from CSV - Export Meta Data to CSV

    Ability to import meta data into selected images from CSV file containing a number of fields, one of which is file name. The other fields would be selected tags to import.
    Import logic would be, Where file name in column 1 matches current selected filename, then take following fields 2, 3, etc and import into Bridge.
    Export selected meta data from selected files to CSV to form reverse of above to allow edit of data in Excel etc.
    Further enchancement on import would be to allow instead of just filename, but date/time range matching.
    Logic would be, Where date/time in column 1 is within X hours of selected
    image creation date/time then import following fields into image meta data.
    Useful for example if you already have a GPS tracklog in CSV and want to try and update images with GPS. Assuming camera clock approx same as GPS clock.

    The issue was within the .csv file.  There were duplicate records that were attempting to load as the ID value.  After deleting the duplicate records, the transformation file ran successfully.
    Debbie

  • SQL bulk copy from csv file - Encoding

    Hi Experts
    This is the first time I am creating a PowerShell script and it is almost working. I just have some problems with the actual bulk import to SQL encoding from the text file since it replaces
    special characters with a question mark. I have set the encoding when creating the csv file but that does not seem to reflect on the actual bulk import. I have tried difference scenarios with the encoding part but I cannot find the proper solution for that.
    To shortly outline what the script does:
    Connect to Active Directory fetching all user - but excluding users in specific OU's
    Export all users to a csv in unicode encoding
    Strip double quote text identifiers (if there is another way of handling that it will be much appreciated)
    Clear all records temporary SQL table
    Import records from csv file to temporary SQL table (this is where the encoding is wrong)
    Update existing records in another table based on the records in the temporary table and insert new record if not found.
    The script looks as the following (any suggestions for optimizing the script are very welcome):
    # CSV file variables
    $path = Split-Path -parent "C:\Temp\ExportADUsers\*.*"
    $filename = "AD_Users.csv"
    $csvfile = $path + "\" + $filename
    $csvdelimiter = ";"
    $firstRowColumns = $true
    # Active Directory variables
    $searchbase = "OU=Users,DC=fabrikam,DC=com"
    $ADServer = 'DC01'
    # Database variables
    $sqlserver = "DB02"
    $database = "My Database"
    $table = "tblADimport"
    $tableEmployee = "tblEmployees"
    # Initialize
    Write-Host "Script started..."
    $elapsed = [System.Diagnostics.Stopwatch]::StartNew()
    # GET DATA FROM ACTIVE DIRECTORY
    # Import the ActiveDirectory Module
    Import-Module ActiveDirectory
    # Get all AD users not in specified OU's
    Write-Host "Retrieving users from Active Directory..."
    $AllADUsers = Get-ADUser -server $ADServer `
    -searchbase $searchbase -Filter * -Properties * |
    ?{$_.DistinguishedName -notmatch 'OU=MeetingRooms,OU=Users,DC=fabrikam,DC=com' `
    -and $_.DistinguishedName -notmatch 'OU=FunctionalMailbox,OU=Users,DC=fabrikam,DC=com'}
    Write-Host "Users retrieved in $($elapsed.Elapsed.ToString())."
    # Define labels and get specific user fields
    Write-Host "Generating CSV file..."
    $AllADUsers |
    Select-Object @{Label = "UNID";Expression = {$_.objectGuid}},
    @{Label = "FirstName";Expression = {$_.GivenName}},
    @{Label = "LastName";Expression = {$_.sn}},
    @{Label = "EmployeeNo";Expression = {$_.EmployeeID}} |
    # Export CSV file and remove text qualifiers
    Export-Csv -NoTypeInformation $csvfile -Encoding Unicode -Delimiter $csvdelimiter
    Write-Host "Removing text qualifiers..."
    (Get-Content $csvfile) | foreach {$_ -replace '"'} | Set-Content $csvfile
    Write-Host "CSV file created in $($elapsed.Elapsed.ToString())."
    # DATABASE IMPORT
    [void][Reflection.Assembly]::LoadWithPartialName("System.Data")
    [void][Reflection.Assembly]::LoadWithPartialName("System.Data.SqlClient")
    $batchsize = 50000
    # Delete all records in AD import table
    Write-Host "Clearing records in AD import table..."
    Invoke-Sqlcmd -Query "DELETE FROM $table" -Database $database -ServerInstance $sqlserver
    # Build the sqlbulkcopy connection, and set the timeout to infinite
    $connectionstring = "Data Source=$sqlserver;Integrated Security=true;Initial Catalog=$database;"
    $bulkcopy = New-Object Data.SqlClient.SqlBulkCopy($connectionstring, [System.Data.SqlClient.SqlBulkCopyOptions]::TableLock)
    $bulkcopy.DestinationTableName = $table
    $bulkcopy.bulkcopyTimeout = 0
    $bulkcopy.batchsize = $batchsize
    # Create the datatable and autogenerate the columns
    $datatable = New-Object System.Data.DataTable
    # Open the text file from disk
    $reader = New-Object System.IO.StreamReader($csvfile)
    $columns = (Get-Content $csvfile -First 1).Split($csvdelimiter)
    if ($firstRowColumns -eq $true) { $null = $reader.readLine()}
    Write-Host "Importing to database..."
    foreach ($column in $columns) {
    $null = $datatable.Columns.Add()
    # Read in the data, line by line
    while (($line = $reader.ReadLine()) -ne $null) {
    $null = $datatable.Rows.Add($line.Split($csvdelimiter))
    $i++; if (($i % $batchsize) -eq 0) {
    $bulkcopy.WriteToServer($datatable)
    Write-Host "$i rows have been inserted in $($elapsed.Elapsed.ToString())."
    $datatable.Clear()
    # Add in all the remaining rows since the last clear
    if($datatable.Rows.Count -gt 0) {
    $bulkcopy.WriteToServer($datatable)
    $datatable.Clear()
    # Clean Up
    Write-Host "CSV file imported in $($elapsed.Elapsed.ToString())."
    $reader.Close(); $reader.Dispose()
    $bulkcopy.Close(); $bulkcopy.Dispose()
    $datatable.Dispose()
    # Sometimes the Garbage Collector takes too long to clear the huge datatable.
    [System.GC]::Collect()
    # Update tblEmployee with imported data
    Write-Host "Updating employee data..."
    $queryUpdateUsers = "UPDATE $($tableEmployee)
    SET $($tableEmployee).EmployeeNumber = $($table).EmployeeNo,
    $($tableEmployee).FirstName = $($table).FirstName,
    $($tableEmployee).LastName = $($table).LastName,
    FROM $($tableEmployee) INNER JOIN $($table) ON $($tableEmployee).UniqueNumber = $($table).UNID
    IF @@ROWCOUNT=0
    INSERT INTO $($tableEmployee) (EmployeeNumber, FirstName, LastName, UniqueNumber)
    SELECT EmployeeNo, FirstName, LastName, UNID
    FROM $($table)"
    try
    Invoke-Sqlcmd -ServerInstance $sqlserver -Database $database -Query $queryUpdateUsers
    Write-Host "Table $($tableEmployee) updated in $($elapsed.Elapsed.ToString())."
    catch
    Write-Host "An error occured when updating $($tableEmployee) $($elapsed.Elapsed.ToString())."
    Write-Host "Script completed in $($elapsed.Elapsed.ToString())."

    I can see that the Export-CSV exports into ANSI though the encoding has been set to UNICODE. Thanks for leading me in the right direction.
    No - it exports as Unicode if set to.
    Your export was wrong and is exporting nothing. Look closely at your code:
    THis line exports nothing in Unicode"
    Export-Csv -NoTypeInformation $csvfile -Encoding Unicode -Delimiter $csvdelimiter
    There is no input object.
    This line converts any file to ansi
    (Get-Content $csvfile) | foreach {$_ -replace '"'} | Set-Content $csvfile
    Set-Content defaults to ANSI so the output file is converted.
    Since you are just dumping into a table by manually building a recorset why not just go direct.  You do not need a CSV.  Just dump theresults of the query to a datatable.
    https://gallery.technet.microsoft.com/scriptcenter/4208a159-a52e-4b99-83d4-8048468d29dd
    This script dumps to a datatable object which can now be used directly in a bulkcopy.
    Here is an example of how easy this is using your script:
    $AllADUsers = Get-ADUser -server $ADServer -searchbase $searchbase -Filter * -Properties GivenName,SN,EmployeeID,objectGUID |
    Where{
    $_.DistinguishedName -notmatch 'OU=MeetingRooms,OU=Users,DC=fabrikam,DC=com'
    -and $_.DistinguishedName -notmatch 'OU=FunctionalMailbox,OU=Users,DC=fabrikam,DC=com'
    } |
    Select-Object @{N='UNID';E={$_.objectGuid}},
    @{N='FirstName';Expression = {$_.GivenName}},
    @{N='LastName';Expression = {$_.sn}},
    @{N=/EmployeeNo;Expression = {$_.EmployeeID}} |
    Out-DataTable
    $AllDUsers is now a datatable.  You can just upload it.
    ¯\_(ツ)_/¯

  • Populating table from csv

    Hi all .. i need the syntax for populating a table from hard drive stored csv file. there are 6 fields in target table (named orgs) and i want to populate its fields 'in' and 'out' from csv file. can someone plz help me? for the time being i m doing it manually by creating scripts in excell sheet but thats not a good approach. i m using oracle 9i. thanx in advance

    Hi,
    Take this example:
    I made a directory 'scott_dir' in my drive & put my csv file in it.
    Then i created an external table using the following:
    create table emp_ext (
    EMPNO NUMBER(4),
    ENAME VARCHAR2(10),
    JOB VARCHAR2(9),
    MGR NUMBER(4),
    HIREDATE DATE,
    SAL NUMBER(7,2),
    COMM NUMBER(7,2),
    DEPTNO NUMBER(2))
    Organization external
    (type oracle_loader
    default directory scott_dir
    access parameters (records delimited by newline
    fields terminated by ',')
    location ('emp_ext.csv'))
    reject limit 1000;
    It will insert data from your csv file to the emp_ext table.
    You can verify data using the Select statement.
    Later on you can merge data in your original table.
    Thanks

  • Unable to synchronise Contacts to BB from CSV file

    I'm trying to upload my contacts from csv file without success. I constantly receive an error message. I'm using
    BB CUrve 8900 and I have a version 6 of Desktop Manager.
    My operating system is Win 7 64bit. I've enabled logging and in the
    Pttrace.log I constatnly see
    "00:14:59.872: Fetching implemented classes at .\iluptbl.cpp line 3408
    00:15:00.523: Fetching implemented classes at .\iluptbl.cpp line 3408
    00:15:02.016: Fetching implemented classes at .\iluptbl.cpp line 3408
    00:15:31.014: Fetching implemented classes at .\iluptbl.cpp line 3408
    00:15:43.051: Fetching implemented classes at .\iluptbl.cpp line 3408
    00:15:43.680: Fetching implemented classes at .\iluptbl.cpp line 3408
    00:15:44.506:
    00:15:44.507: Begin ILX Session, Source=Importer/Eksporter ASCII, Target=Urządzenie
    00:15:44.531:
    00:15:44.532: Starting One-Way Sync (from scratch)
    00:15:44.803: Phase=10, User=22E867D2: Reading data from Urządzenie Książka adresowa
    00:15:45.111: Finished Reading 17 Records + 0 Deletes + 0 Unchangeds for User=22E867D2 from Urządzenie Książka adresowa (slow sync input for resync)
    00:15:45.112: Phase=20, User=22E867D2: Reading data from Importer/Eksporter ASCII Książka adresowa
    Tue Apr 26 00:15:45 2011: Error Blad interfejsu API..4238 at .\AsciiConnectorIConnector.cpp line 375
    Tue Apr 26 00:15:45 2011: Error 4107.4107 at .\sdk_data.cpp line 875
    Tue Apr 26 00:15:45 2011: Error 0.4107 at .\ciltrans.cpp line 241
    Tue Apr 26 00:15:45 2011: Error 4107.4107 at .\Ilx_sdk.cpp line 220
    00:15:45.134: Translation Unit Status: User=22E867D2, rc=87, Phase=20, TrErr=4107, SysErr=0 at .\xlatev3.cpp line 650
    00:15:45.136:
    00:15:45.137: End ILX Session, elapsed time =   0.000 seconds
    In Tif.log i see that program is successfully exporting data from device to tif file but when it attempts to open
    csv file it fails
    "--------  ilsdk finished processing  a_record_for_export_to_TIF, rc=0, nRc=0, action=19
     --------  ilsdk starting assembly of a_record_for_export_to_TIF
     --------  END OF FILE
     --------  ilsdk finished processing  a_record_for_export_to_TIF, rc=0, nRc=4006, action=3
     --------  Ending Load From Target phase; starting next phase
    Finished Reading 17 Records + 0 Deletes + 0 Unchangeds for User=22E867D2 from Urządzenie Książka adresowa (slow sync input for resync)
    ILTIFReopenFile/NO-OP
     --------  Ending next phase; starting Load From Source phase
    Could you help me as entering over 600 contacts would be painfull
    brgds
    WIesiek

    This is exactly what I am dealing with. Blackberry Link will work great for awhile and then something happens and I can not sync my contacts. I have spent many hours and days trying to fix this to no avail. BB support has been great but I still go through a long process each time. This week I finally gave up and started using the Microsoft Hotmail Outlook Connector since my company does not offer Active Sync or Exchange. It works flawlessly. I would dedfinitely check it out. If you swith to a Widows Phone or Android this is one of the options they recommend you use.
    In a nutshell Link is awful and BB should be ashamed of the product they put on the market. I will only use it for syncing pictures and documents. Good luck!

  • How to import data from CSV to TCode COR6N

    Hi friends,
    Please help me. I don't know how to import data from CSV file to TCode COR6N. (time ticket for production order)
    I used function: BAPI_PRODORDCONF_CREATE_TT to fill information into TCode COR6n, but it thrown error: "Orders of the order type ZP01 cannot be confirmed". ZP01 is my Order Type.
    Thanks in advance.
    Nguyen Phuoc Toan.

    You can use the LSMW tool to do this.
    The details of LSMW BAPI method is as follows :-
    Business Object:- BUS2116
    Method:-CREATEPREDEFTIMETICKETMULTIPLE
    Message Type:- PPCC2PRETTICKET
    Basic Type:- PPCC2PRETTICKET01
    Regards
    Dhirendra

  • Determine propagation time from signals obtained with Gould Transition software

    Hi
    I'm trying to analyze signals from measurements of sound waves. I tried to do it with the Transition software, but this software is not sufficient because there are no averaging or filter options (furthermore you have to analyze graph after graph). Signals are saved in ASCII-code and what I'm looking for is a good and fast way to determine propagation time from the differences of the arriving of two waves. FFT would be great to.
    Thanks a lot
    Marco

    Any personal experiences would be greatly appreciated.

Maybe you are looking for