Data Acquisition form single precision to binary format

Hi everybody,
I work in an Astrophysics group and I'm developing an acquisition system of signals generated by X-Ray detectors (Transition Edge Sensor).
I'm trying to save data acquired by NI-PXI 6132 (DAQ-mx software in LabVIEW 8.0) in binary format.
To do this, I use the 'Write To Measurement File' procedure.
I don't know how it is possible to change the precision from the DBL to SGL directly from DAQ-mx, but I solve this point by creating a 'For Loop' inside my VI, after the data is acquired.
The problem is that I try to save the SGL data with the previous procedure but the 'Convert To Dynamic Data' automatically occurs, and the data are saved in binary format but in doube precision. So, the previous conversion that I made from DBL to SGL does not work.
How can I solve this problem?
Thank you for your help!
Ciao,
Claudio!

Hi Alessio,
thank you for your prompt reply!
In attach there is the VI you reqeusted me.
Ciao,
Claudio
Attachments:
Daq_2monitor_BINARY_ASCII_single_float.vi ‏498 KB

Similar Messages

  • Data Acqusition and recording in Binary Format

    Hi
    Can anyone tell me how to perform buffered Data acquisition and write
    it in Binary format.
    I want to record data when it is streaming that is i would like to
    record data on a real time basis.
    I have started acquiring data and also graphed it, But it is not a
    buffered acquisition, Can i record this data in the binary format??
    Srikanth

    Hi,
    Labview has an example that does exactly what you are asking for:
    Cont Acq to File (binary).vi
    You'll find it in:
    \examples\daq\anlogin\strmdisk.llb
    Good luck,
    Alberto

  • Extract FORM data into BINARY format or XSTRING format?

    Hello,
    Our config: I am developing an VISITOR Adobe Interactive Form, i have Designer 8.1 and Reader 9.0 in my Windows 7 Laptop. When i checked the menu EDIT-->form properties --> Compatability, its Adobe XML form File (XDP)
    When i opened this VISITOR interactive form in a browser, right clicking and checked form PROPERTIES, then i saw the "PDF Version: 1.6 (Acrobat 7.x)"
    I am using JavaScript
    1st user opened the form, entered data, say entered USA in COUNTRY field.
    2nd user opened the same form and edited the COUNTRY field from USA to UK.
    Pls. let me know is there any way for me to capture the user entered data into a BINARY format or XSTRING or RAW FORMAT or RAW STRIGN
    Pls. note i need latest changes, like COUNTRY field hold the value of UK.
    Pls. note for some reason, SAVE the form should not come into picture in my case (form will be saved in some workflow (pls. note not adobe's workflow, we hv SAP workflow) staging area in the time lapse of between  1st user and 2nd user,)
    Thank you

    I am looking WHOLE FORMA DATA in either XSTRING or BINARY FORMAT or any other format, except STRING format.
    I gave COUNTRY as an example, but am looking the HWOLE form AS IS
    Any help pls.
    Thank you   

  • Is BPEL supports writing character data into binary format  ?

    Hello,
    I can write a BLOB data from DB to a file using Opaque feature in File Adapter.
    But is it possible to write Character data into Binary or Hex Format either by using base64Binary or Hex or bytes formats supported in XML?
    Thanks

    I am looking WHOLE FORMA DATA in either XSTRING or BINARY FORMAT or any other format, except STRING format.
    I gave COUNTRY as an example, but am looking the HWOLE form AS IS
    Any help pls.
    Thank you   

  • Storing analog output data in a datatype other than single-precision

    I'm using three PCI-6713 AO cards in a single computer to output 24 channels of data using Labview 6.1. The input terminals of all the analog output VI's are of type single precision (32-bit). The output cards only have 12-bit DAC's, so there is a large amount of information stored that is never used by the cards. Is there a way in Labview to store AO data in another, smaller datatype - such as a 16-bit integer. This would greatly reduce the memory requirements for my application.

    I just figured out the answer to my own question. There isn't an option for binary in the AO Write vi, but if you go down to the AO Buffer Write vi (which is called by AO Write), you can select binary array.
    Thanks for all your help.
    Josh

  • Extract Form either in BINARY FORMAT or XSTRING?

    Hello,
    Our config: I am developing an VISITOR Adobe Interactive Form, i have Designer 8.1 and Reader 9.0 in my Windows 7 Laptop. When i checked the menu EDIT-->form properties --> Compatability, its Adobe XML form File (XDP)
    When i opened this VISITOR interactive form in a browser, right clicking and checked form PROPERTIES, then i saw the "PDF Version: 1.6 (Acrobat 7.x)"
    I am using JavaScript
    1st user opened the form, entered data, say entered USA in COUNTRY field.
    2nd user opened the same form and edited the COUNTRY field from USA to UK.
    Pls. let me know is there any way for me to capture the user entered data into a BINARY format or XSTRING or RAW FORMAT or RAW STRIGN
    Pls. note i need latest changes, like COUNTRY field hold the value of UK.
    Pls. note for some reason, SAVE the form should not come into picture in my case (form will be saved in some workflow (pls. note not adobe's workflow, we hv SAP workflow) staging area in the time lapse of between  1st user and 2nd user,)
    Thank you

    I am looking WHOLE FORMA DATA in either XSTRING or BINARY FORMAT or any other format, except STRING format.
    I gave COUNTRY as an example, but am looking the HWOLE form AS IS
    Any help pls.
    Thank you   

  • Binary data to form a character

    Hi!
    I'm making a byte array from reading a file.
    Could anyone tell me how can I achieve 3 bytes of binary data to form a 6 character number?

    So I have a binary file so called "Input file" . I' ve also got the "Output file", that has similar lines like this:
    IPX5  74002709       0312359203 133586 0,004   CU.1.IPX0My job is to find out how I can get the output look from the input binary file according to a specification...
    The specification says that there are 6 parts(these are the separated parts of the output lines) which are repeated in the binary Input file.
    Two parts of them are ASCII data: the first part and the last part of the output line.
    I've managed to read these from the input file using a byte array that I've converted into char, so I've got the first and last parts as output.
    But the other 4 parts aren't ASCII data but character numbers.These have been converted into char,too, but I can't see them in the output..
    For these parts the specification tells that for example: there is a part that is contained in 3 bytes and extracted to form a 6 character number...
    So this is the point what I don' t get it how to solve this "byte to number" conversion...
    I hope it would be clear enough..

  • Single Template support multiple formats of data - Best practice

    I have a requirement to create Invoices in a single PDF file.
    The invoices would belong to different categories - DebitNotes, CreditNotes, Invoices with single product,
    invoices with multiple product etc.. each will have a different format.
    I initially thought the right way to create a single pdf is to use a
    single Template, with different formats of invoice seperated by conditional formatting.
    The see from reading the blogs that the other way is to create sub-template
    (one each for credit, invoice, Debits etc) and plug it into the
    main template.
    I would like to know what is the best practice that is followed in the above case.
    If I were to use sub-templates how would I make it possible to view the invoice stub only on the first page.
    Since the data from the sub-template would go to multiple pages.
    Is adding the stub data to the footer the only option. Please can someone share with me an example
    template.
    Thanks
    Shandrila

    Shandrila
    If the various document types are of a single XML format ie the same structure with just document type differences and the layout format is the same just the data different or very minimal changes that can be handled with conditional formatting then I think it would be OK to have a single report for all document types.
    If the data structures are very different and the layout requirements are different then I would create separate reports for each document type. If the data structure is teh same but the document type layouts are different then go for separate layout formats.
    Going down the sub template path can be a little difficult, you might end up with a very complex set of templates that are almost as much of a pain to manage as the original report you are trying to replace.
    Here's the best scenario IMHO ...
    1 data extract, parameterized to pull invoice, CM, etc data based on the user request
    Multiple layout templates, 1 for each document type. If you have common layout sections across the layouts e.g address blocks then break them out as sub template components that all of the layouts can access and share.
    Multiple report definitions, sharing the data extract with a single layout template associated with them
    cheers
    Tim

  • Exporting form data to a single table

    I have created a form in LiveCycle Designer ES 8.2  and simply want to export the data into a single table that I can then import into Access. It appears that in LiveCycle the row elements of a table are children of the table element and the column elements are children of the row elements.  I am wondering if there is an easy way to design a form in LiveCycle so that each cell of a table is a root element like the rest of the fields in the form.  I basically want to export one row of data from each form and append it to a larger table. Exporting the data as csv or tab-delimited would be sufficient, but neither of these appears to be an option.

    The only way I've found to do this is to open the form in Acrobat, click Forms, mouse over Manage Form Data, then click Export Data. Your only options are to save as .xml or .xdp. I saved my data as .xml. I was then able to import the .xml file into excel by clicking Data, mousing over xml, clicking Import... and browsing to the Acrobat generated .xml file. I know this is convoluted but it's the only way I've found to get the data into a spreadsheet. Acrobat exports the data as filename_data.xml, i.e. a form named contacts will export it's data as contacts_data.xml. Be aware, I've only tested this on a form that had a table in it and nothing else. If your table is part of a form with other fields in it I have no idea how the other data will look after exporting.

  • Importing data in obscure binary format into diadem

    Hi all,
    Lets assume that I have some data in a binary format which DIAdem can not automatically import into the data portal. One method I am aware of and used is to create a dataplugin in vbs which will then allow me to import this data. Recently I came across something called the "TDM header writer" which gives me the impression that it is capable of creating a header file for my binary data and as a result will be able to import this data into DIAdem without creating a dataplugin. Is this actually the case? If the answer is yes which method is actually easiest to use dataplugin or TDM header writer?
    Thanks for reading and your assistance in advance.
    Matthew

    Hi,
    your impression is correct. With the TDM Header Writer you can write a TDM header file for your binary data, so it can be imported in DIAdem as TDM file.
    However, you will need to create a TDM Header for each of your binary files. There are several ways to use the TDM Header Writer:
    - C API
    - JAVA API
    - LabVIEW VIs
    But please let me ask, if you are not satisfied with the DataPlugin you created. It should enable you to read all of your binary files without any extra work. You can export DataPlugins and distribute them. So once created, it should feel like a native file format for DIAdem.
    For an overview of DataPlugins provided by National Instruments, please have a look here:
    DataPlugins

  • Data Acquisition - using local variables to write data to a file

    Hello,
    I am running a Data Acquisition vi (currently in LabVIEW 7.1 but soon to be updated to 8.2) that collects ~100 parameters of data from several sources contained in a while loop. The current configuration (which I did not write) uses very few subVIs and writes to ~100 local variables to store each parameter. It then reads all the local variables and builds an array of all the strings, converts then to a spreadsheet string, then uses the write characters to file function to append to a datafile. I am trying to clean things up and have came up with subVIs to collect the data from the following sources:
    8 serial port sources collecting btwn 8 and 20 parameters each
    ~15 thermocouple readings
    ~10 analog inputs
    ~20 parameters read off an ARINC 429 bus.
    I have come up with a subVI to read each of the sources and have placed the subVIs in the while loop. Each subVI outputs the data that it collects in array or cluster form. I was wondering how best to write each parameter to a CSV file at between 1 and 10 Hz. Should I write each subVI output to a LV and then read them off as was done before (the difference being that I have reduced the # of LVs to ~10 vs >100?
    I should add that precise timing is not that important, so if all the subVIs are not collecting simultaneously (which I understand that they won't be), it does not really matter.
    Thanks.

    Hi jilla,
    jilla wrote:
    What I think that you are saying is to turn the outputs of the 4 subVIs into inputs of a 5th subVI that writes to the data file. Correct?
    Yes.  It may sound like a fine-point, but I beileve it's better to create a VI specifically for formatting data - in your example, 4 arrays IN, a single string OUT.  Then write the string to file as a seperate operation.  GUI-displayed data can go through a similar transformation, the four arrays wired to a subVI which builds output-structures specifically for display.  It's a beginner's mistake to put lots of individual controls and indicators on the screen when groups of them are naturally related (in an object-oriented sense.)  Use clusters to group related controls - this will keep the diagram much cleaner.
    One more question: at what point (either # of data points or frequency of data collection) does it become necessary to use queues? Thanks.
    Well, there's not really a clearly definable "point".  I'd say if your update-rate climbs above 100Hz, or you witness poor program or system performance, then it's time.  The scenario you've described is a fairly simple acquire/display&log loop - and simple is good.   Then-again people can't see/react-to updates faster than about 10Hz - so it doesn't make sense to sacrifice performance - if performance becomes an issue.
    Re: queues:  Queues are sometimes used to buffer data that's "produced" in one place and "consumed" in another.
    Here, if/when logging data, you're logging with every DAQ.  I wouldn't recommend using a queue to transport data from a "DAQ loop" to a "Logging-loop" - those functions can be in the same loop.  Should/could a queue be used to get data from a "DAQ loop" to update the GUI at a lower frequency?  Sure, but a Notifier might be a better choice.   Further, in the (simple?) program you've described, you might use a case structure (True/False) to only update FP indicators every "X" iterations - a simple solution that doesn't require Queues or Notifiers.
    Cheers!
    "Inside every large program is a small program struggling to get out." (attributed to Tony Hoare)

  • Single precision conversion

    Can someone explain why doing a floating point conversion compromises digit of precision accuracy? For example, if I wire the double precision number "16.0001000" to the "to Single Precision Float" vi input, the output is "16.0000991". 
    I have an application in which this level of accuracy is critical. I can think of several work arounds, but I still would like a good explanation for why this is the case.
    pmac

    Just use array to spreadsheet string on the DBL and write as text file (and an analogous operation for reading, converting back to DBL). Alternatively, you could make DBL version of "read/write  spreadheet file" by editing them slightly and saving them under a new name elsewhere. Make sure to change the icon too, e.g. give it a different color.
    If precision is that important, you might want to use binary files. The don't have any formatting loss at all.
    Of course if this is just measurement data, don't be fooled by fake precision. even SGL has a 22 bit mantissa, so if your DAQ hardware acqures in 16 bits, SGL is plenty.
    LabVIEW Champion . Do more with less code and in less time .

  • Data acquisition with highly accurate GPS time stamping (hh:mm:ss:sss UT)

    A research group has developed a data acquisition system dedicated to daily observations of solar phenomena at radio waves using high time resolution (5-100 ms).
    First, the analog signal from the antenna (1000-2500 MHz) is sent to a spectrum analyzer (HP8559-A). After that, the analog signal is digitized during 5 minutes in a NI USB-6009 device connected to the PC, before to be stored in a file. By that time, other acquisition data cycle is restarted and so on up to the end of day.
    However, we need timestamp (UTC) the solar data at the beginning and end of each sample of 5 milliseconds, during all acquisition process, in order to have the UTC time information of recorded solar events. So, a time resolution of at least 1 ms is required.
    A software permit us to read recorded data and visualize corresponding spectra during last 5 minutes. However, we need UTC time information in the format hh:mm:ss.sss UT. Then, UTC millisecond is important for us, but not available yet.
    To solve this problem, we are searching for a high accuracy timing GPS receiver (on the order of microseconds or nanoseconds), with at least three significant digits of precision (UT time in milliseconds) and which give us the UTC signal at least each 5 milliseconds or even better (1 ms). However, most GPS receivers only outputs 1 PPS (Pulse Per Second = 1000 millisecond).
    Then, we are open to a better addressing on how to solve this problem !!
    We know that GPS time synchronization with PXI-6608 can be used to precisely timestamp data acquisition, but we have to solve this problem with a lower cost. Also, we require a PCI interfacing instead of PXI.
    Do we have to use a timer card? How do we use a timer, GPS and USB-6009 to timestamp data acquisition? Does the NI PCI-6602 solve the problem?
    Obs:
    The data acquisition software was developed in C++ Builder to run on a PC (Pentium III 2 GHz) under Windows XP environment. We used the NI-DAQmxBase driver to sample the signal from the NI USB-6009.
    We are sure your expertise can help us on this matter.
    Regards,

    I better explain my problem in the thread How to timestamp continuous analog data acquisition using USB-6009 and GPS timing board?
    We are considering to use the PCI-1588 or a GPS timing board (low-profile slot too), but I´m not sure that the USB-6009 can be used to solve the problem.
    How to syncronize the data acquisition and timestamps using USB-6009 and PCI-1588 with external GPS?
    In other thread I would appreciate any suggestions for GPS timing board and PCI-1588 too.
    Regards,
    Lilian

  • How to store measurement data in a single database row

    I have to store time-data series into a database and do some analysis using Matlab later. Therefore the question might be more a database question rather than Diadem itself. Nevertheless I'm interested if anyone has some best practices for this issue.
    I have a database which holds lifecycle records for certain components of same nature and different variants. Depending on the variant I have test setups which record certain data. There is a common set of component properties and a varying number of environmental measurements to be included. Also the duration of data acquisition varies across the variants.
    Therefore having tables appears to be non-optimal solution for storing the data because the needed number of columns is unknown. Additionally I cannot create individual tables for each sample of a variant. This would produce to many tables.
    So there are different approaches I have thought of:
    Saving the TDM and TDX files as text respectively as BLOB
    This makes it necessary to use intermediate files.
    Saving the data as XML text
    I don’t know yet if I can process XML data in Matlab
    Has anybody an advice on that problem?
    Regards
    Chris

    Chris
    Sorry for the lateness in replying to your post. 
    I have done quite a bit of using a Database to store test results.  (In my case this was Oracle DB, using a system called ASAM-ODS)
    My 2 Cents:
    Three functions were needed by users for me.  1) To search and find the tests,  and  2)  To take the list of Tests and process the data into a report/output summary. 2) If the file size is large, then being able to import the data quickly into analysis tool speeds up processing.
    1) Searching for test results.  This all depends on what parameters are of value for searching.  In practice this is a smaller list of values(usually under 20), but I have had great difficulty getting users to agree on what these parameters are. They tend to want to search for anything.   The organization of the searching parameters has direct relationship to how you can search.   The NI Datafinder does a nice job of searching for parameters, so long as the parameter data are stored in properties of Channel Groups or Channels. It does not search or index if the values are in channel values.
    Another note: Given these are searchable parameters, it helps greatly if these parameters have a controlled entry, so that the parameters are consistent over all tests, and not dependent on free form entry by each operator. Searching becomes impossible if the operators enter dates/ names in wildly different formats.
    2) Similar issue exists if put values into databases. (I will use the database terms of Table and column(Parameter) and Row (instance of Data that would be one test record.)
    The sql select statement, can do fast finds, if store the searchable parameters in rows of a table. Where would have one row for each Test Record.   The files I worked with have more than 2000 parameters.   Making a table that would store all of these, and be able to search for all of these, makes a very large number of rows. I did not like this approach, as it has substantial maintenance time, and when changes are made, things get ugly quick.
    3)This is where having the file format be something that the analysis tool can quickly load up is beneficial.   Especially if the data files are large. In DIAdem's case, it reads TDM,TDMS files very quickly into the dataportal.   It can also read in the MDF or HDF files, but these are hierarchical structures that require custom code to traverse the information, and get the information into dataportal to use in Analysis /reporting. (Takes more time to read data, but have much more flexibility in the data structure than the two level tdm/tdms format.)
    My personal preferences
    I would not want to put the test data into a Table row. Each of the columns would be fixed and the table would be very wide in columns.
    >
    I personally like to put the test Data into a file, like TDMS, MDF, or HDF and then the database would have a entry for the reference to the attachment. The values that are in the database is just the parameters that are used for test Searching, either in datafinder or in sql commands in the user interface.
    Hopefully these comments help your tasks some.
    Respectfully,
    Paul
    tdmsmith.com

  • Data Acquisition on a 20MHz Timebase with PCI-6110

    I would like to acquire a data point (viz a voltage value) every 50ns (20MHz timebase). Then I would like to write these voltage values and the corresponding time value to a text file and finally plot the voltage values over the time. The data acquisition should be started on a trigger. How can I program this application in LabVIEW 7.1.1? I use a PCI-6110 card for the data acquisition. Thanks for any help!

    I would suggest you look at the shipping examples. Open the Example Finder and from the browse tab, select Hardware Input and Output>DAQmx, Analog Measurements>Voltage. Start with one of the simpler examples such as Cont Acq&Graph Voltage-Int Clk. You can also try the DAQ Assistant. Also, saving data to a text file can be pretty slow and will be a problem trying to maintain high acquisition rates. If you acquire data for a finite amount of time, you'll probably want to save the data once it's all been acquired and not real-time. You can use any number of methods for this. The Write LabVIEW Measurements File can be used to save an array of waveforms that you've captured. There are other shipping examples for saving real-time in a binary format. There are other examples at the NI Developer Zone.

Maybe you are looking for

  • When saving a page or link the name of the file should be the folder name, not the number in the address line

    Vhttp://www.commondreams.org/view/2010/08/07-2 should be something like What Collapsing Empire Looks Like I want to File> save page as or File >save Frame as I get something like, " 07-2.htm" how do I get the better name for the saved file?

  • Current DDR Clock

    Greetings - My Fellow MSI users!! This is my first post and I'd like to welcome myself...   Thank you, thank you!! I must start off with saying that I've built a number of rigs before, but to date I'm having the most fun with this one (see specs in s

  • Problem connecting weblogic OSB with IBM websphere via foreign JMS

    Hi All, I am trying to setup Foreign JMS. My configuration doesn't work. Could you please let me know if I am missing anything . All my configuration details are done as specified in this link http://www.oracle.com/technology/products/integration/ser

  • Alot of files missing

    1 day i woke up and i couldnt use any programs that used JRE i looked and my java folder was pretty much empty.. i didnt know what to do i tried to download but it said that its installed already and bla bla... so i went on my sisters computer copied

  • "ZENNW32.DLL is misssing the entry point nwappLoadOD."

    We are getting the error "ZENNW32.DLL is misssing the entry point nwappLoadOD." I have found this in some other threads in this forum, but they did not go in a useful direction for me. This error message is appearing on a WinXP Pro workstation before