Signal express tc calibratio​n to daq assistant calibratio​n

I ran some calibrations in Signal Express for K-type thermocouples. I didn't realize that those calibrations don't carry over to other NI programs. I would like those same calibration values to be used in the DAQ Assistant, rather than completing all my calibrations over again. Is there a way that LabVIEW can grab the Signal Express calibrations and use them?

Hi Jdezman,
I will try and help you with this aspect of hardware calibration.
NI-DAQmx will allow you to do a system calibration. Once creating your task in signal express, there will be a 'Calibration' tab under the 'Thermocouple Setup'. This will direct you to the 'Calibration Wizard'. This Wizard is the same in both Signal Express and the 'Measurement and Automation Explorer' (MAX).
There is a KnowledgeBased article which will guide you through this process:
http://zone.ni.com/devzone/cda/tut/p/id/4224
Hope this helps with your system calibration. Please let me know how you get on,
Regards,
Aaron. E
Applications Engineer Team Lead
National Instruments
ni.com/support

Similar Messages

  • How to select signals from the output of a DAQ assistant

    Hi! 
    I am a new user of Labview 8.5 and I am working with a USB-6210. I have two different instruments connected to the same USB device, half of the channels are being used for pressure transducers where I only need to reed and save the data, while the other half are associated to TCD detectors where I need to perform some analysis to the signal to obtain and save the data. I am in the first step of building the block diagram, After I define each of the signals which its corresponding channel using the DAQ assistant, I need to select and separate the signals coming from the pressure transducers from those coming from the TCD detectors, before I can continue drawing the block diagram. I am trying to use the Select Signal function but I am not sure how to do this. Can anyone please let me know at least in which manual can I find a good explanation? I have been reading the Getting Started and the User Manual for LabView but they have not been very useful so far.
    Thanks!
    Solved!
    Go to Solution.

    Wire the Select Signals express VI to your DAQ Assistant and run the VI once (this allows the signal names to propegate).  Then when you open the Select Signals configuration you'll be able to pick out the channels you want by name instead of just "Signal 1" etc.
    It sounds like you'll probably want two Select Signal Express VIs, one for the TCD detectors and one for the pressure transducers.  Something like this:
    If you right click on the VI and select Help or click on the Help button within the configuration dialog you will bring up the LabVIEW help page for this (or any other) VI.  Are you looking for some other information that isn't documented in the help file?
    Hope this helps,
    Simon

  • In Labview 8.5, what happens if the signal input exceeds the signal input range set by the DAQ Assistant?

    Hello all,
    This should be a pretty simple question, but I can't seem to find the answer on-line and don't currently have the capabilities to test this:
    I'm using LabVIEW 8.5 and have a VI that imports sensor data through the DAQ Assistant. In the configuration tab there is a signal input range. What happens if my sensor exceeds this range? Will I get a warning? Will the value default to the maximum (or minimum)? I was interested in writing in some code to display an error as I approach the limits of this range, but was unsure if I also needed to include some code to display an error if the range is exceeded as well.
    Thanks for the help,
    Tristan
    Solved!
    Go to Solution.

    Hello Tristan,
    The behavior depends on the range you choose and the device you are using.
    If you are using a device with only one valid input range, we will use this range even if you set a smaller minimum and maximum in the DAQ Assistant.  Thus, if your device only supports ±10V and you set the range to ±8V, you will still continue to get valid data after your sensor exceeds 8V until you approach 10V.  Once you reach the limit of the range of your device, the output will "rail" and just return the maximum value until the signal drops below the maximum again.
    Note: A device that is nominally ±10V usually has some overshoot (like ±10.2V) that is typically specced in the manual.
    However, if you are using a device with multiple input ranges then things get more complex.
    The NI-DAQmx drive will pick the smallest range that fully encompasses the range you choose.  So, suppose your device supports the following input ranges: ±0.2V, ±1, ±5V, and ±10V and you choose 0V - 3V as the range in the DAQ assistant.  The NI-DAQmx driver is going to look at your input range and the list of input ranges that your hardware supports and choose the smallest that encompasses the full range you set.  This would the ±5V, because that's the only range that contains up to 3V.  As a result, any input signal between ±5V will be returned and any outside this range will "rail" to either the maximum or minimum value.
    We do this because using smaller ranges make more effective use of the resolution of the ADC.  Thus we try to use the most efficient range based on what you request without picking a range that will cause you to miss data.
    Let me know if I can clarify this further. 
    Seth B.
    Staff Test Engineer | National Instruments
    Certified LabVIEW Developer
    Certified TestStand Developer
    “Engineers like to solve problems. If there are no problems handily available, they will create their own problems.”- Scott Adams

  • Linear fit between two cursors from a signal from a NI-instruments via DAQ-Assistant

    Hi ! 
    I'm currently having som problem with my labview code. I'm trying to make regression on a specific part of my graph. The graph consist of a signal gathered from a NI cDAQ-9178. Currently I can get the live data in a graph (Raw Signal) and then  "zoom in" to the part of the graph that im intrested in by two cursors and then display the results in a new graph (Adjusted Signal). Now a want to make regresion in the new graph. How can i do this the best way ? 
    To be noted I'm a beginner of labview so my code maybe a bit off.
    Best regards Maurlind
    Solved!
    Go to Solution.
    Attachments:
    Rawsignal test.vi ‏245 KB
    rawsignal test.png ‏64 KB

    Hi maurlind,
    I got it, the function that derived the linear fit couldn't handle that large numbers on the x-axis. I edited the VI so that the large number is suctracted for the fit, i.e. the array will now begin at zero time for the fit. It will not affect your plots though, only the algorithm .
    Try it now with the files I provided. The file "pm_mod2_Rawsignal test.vi" should be used with "test3.lvm" (your data). The file "pm_mod3_Rawsignal test.vi" is the same as "pm_mod2_Rawsignal test.vi", but it is arranged to work with the simulated data. The file "test3.lvm" is your data, but the file is slightly edited to work on my system.
    Edit "pm_mod3_Rawsignal test.vi" to fit with your acquired signal as you did with the first file you got from me.
    Attachments:
    1412794.zip ‏483 KB

  • How do I use DAQ Assistant to output DC on an AO?

    My background includes 15 years of programming PLC's. AO's were simple. You passed the AO a value between preset ranges and it output a proportional voltage signal based on the hardware capabilities.
    I foolishly thought LabView and DAQ AO's would be about the same. I built a vi to generate a DC signal which I could control the amplitude of. I also tied it to a chart to monitor how it worked. I managed to draw some very pretty ramping pictures on my chart, but when I put a DMM across the physical terminals I see no voltage.
    Then I used MAX to test the channel and measured it again. Now I see voltage on my DMM. I was attempting to use the DAQ Assistant to do this. Do I need to force it to re-read the changing value I am sending it? All of the examples I find show me more primitive methods, and I could not seem to get any of them to do what I want either. I need to send a speed reference to a DC drive.
    My incorrect method was to recursively add a value for ramping the signal up or subtract to ramp down. This method works perfectly on an industrial PLC. I connected a chart to the signal just as it entered the DAQ Assistant, but the only way I could check the actual was with the DMM.
    Are there any examples that use the Assistant? If not then I guess I will have figure out how to use the primitives anyway.
    technomage

    Hello technomage,
    It looks like we have discussed this problem in another thread (found here).  If you have any other questions about it, please post there.
    Regards,
    Jesse O.
    Applications Engineering
    National Instruments
    Jesse O. | National Instruments R&D

  • Where can i find labview signal express DAQ assistant at?

    Hi, i need help
    Q1) where can i locate Labview signal express DAQ assistant at?
    Q2) Is Labview signal express functions same as Labview 8.6 ?

    LabVIEW Signal Express and LabVIEW are two different products.
    If you are talking about the DAQ Assistant, then you need to have the latest DAQmx driver installed and you should be able to find in the Measurement I/O palette on the LabVIEW Block Diagram.
    Adnan Zafar
    Certified LabVIEW Architect
    Coleman Technologies

  • Missing wire connectors on DAQ Assistant express VI's

    I'm seeing some strange behavior on one specific computer running LabVIEW. The problem is that DAQ Assistant VI's have no connector points. I have nine other identical computers (used in a classroom setting) but I don't see this behavior on the others (I haven't tested them all).
    I've removed LabVIEW (and all other NI software) and re-installed it (LabVIEW v 8.6.1). I also tried a more recent version of LabVIEW (2010?). Same behavior, no connector dots on DAQ Assistant VI's.
    The computers use DAQmx with USB hardware (cDAQ).
    I guess my next step is to format the hard drive and re-install everything. My theory is there's some kind of driver corruption. What do you think?

    Allie,
    Here are the steps I use to create the Express VI using the DAQ Assistant:
    * On the block diagram:
    Right click -> Measurement I/O -> MX -> DAQ Assistant
    * Drop Blue box on block diagram.
    * Configuration Screen Appears
    Aqcuire Signals -> Analog Input -> Voltage -> Dev1 (USB 9219) -> ai0 -> Finish
    * Voltage Input Setup Appears
    Accept default settings -> click OK
    * Voltage Input screen disappears, a blue DAQ Assistant appears on block diagram. No connection points.
    When I execute the exact same procedure on another computer I get a DAQ Assistant with connection points.
    Weird eh?
    thanks!

  • Mutiple readings from Express DAQ Assistant VI

    Hello All!
    First I'm a fresh noob to LabView and although I understand most of the basic concepts, I'm also new to programming in general. I'm working on a project to remotely monitor several different parameters of a system. Right now I'm working on a test program to monitor three parameters in my shop across the street. I'm using a NI USB-6008 as a DAQ device, DAQmx 8.0 and LabView 8.0. All of my input signals are 4-20 mA signals coming from my three sensing devices. 
    Here's my problem; I need to monitor all three signals on digital displays. I can accomplish this by using three different instances of the express DAQ assistant but then I get the dreaded "50103 error" unless I wire the stopped output of one assistant to the stop input of the next in a feedback loop and even then I get the 50103 error on the first iteration of the loop. From my research I understand that this isn't the correct way to program this. I should use one DAQ assistant with multiple channels assigned to each one of my parameters. I've did this but now how do I make each one read out on an individual digital display? I only have one output from my DAQ assistant.

    Hi George,
    Use split signal in the signal manipulation functions menu to get your 3 individual singals on respective indicators
    Regards
    Dev
    Attachments:
    SPLIT.vi ‏85 KB

  • Can't see the "DAQ assistant express VI" in LabView7 express

    i am using the LabView 7 express avaluation package, and trying to use my PCI 6132 DAQ card with it.
    i cannot see the "DAQ assistant express VI" simbol in the "input" box, and therefore cannot connect to the DAQ.
    all drivers are installed, and i manage to log data directly from the DAQ, using the "Measurement & Automation" software.
    any ideas?
    doe's the evaluation version even supports this DAQ?

    GilKlar,
    Did you install DAQ first, or LabVIEW? The most common cause of this problem is that DAQ support was installed before LabVIEW support. In this case, the DAQ driver will not find any valid installations of LabVIEW and thus, will not install the necessary VIs. Please try repairing the DAQ installation and making sure that support for LabVIEW is checked.
    If this does not solve your problem, please let us know what version of the DAQ driver you are using, as well as how you are acquiring data in MAX.
    Hope this helps,
    Ryan Verret
    Product Marketing Engineer
    Signal Generators
    National Instruments

  • I have a DAQ Assistant configured to read multiple channels at the same time. When I wire a graph indicator to the output, I see all of my signals jumbled together. How do I split them up into seperate signals?

    I have a DAQ Assistant configured to read 2 channels at the same
    time. When I wire a graph indicator to the output, I see the 2
    signals jumbled together. How do I split them up into seperate signals?
    When I wire any type of indicator it is showing just one output of a single channel.
    I want 2 indicators showing 2 different signals as expected from the 2 channels configured. How to do this?
    I have tried using split signal but it end up showing only 1 output from 1 signal in both the indicators.
    thanks in advance.
    Solved!
    Go to Solution.

    Yes you are right. I tried that but I did not get the result.
    I just found the way. When we launch split signal, we should expand it (split signal icon) from above and not from below. It took me a while to figure out this. 
    thanks 

  • How to control a digital output signal using the DAQ assistant?

    I am using a USB 6251 DAQ board and would like to control a switch. I am gathering airflow, pressure, and acoustic data on the same board (analog input) and would like the switch to trigger when the airflow value is within a pre-determined range. Finally, I would like to reverse the polarity of the switch (to off) about a second after the initial digital signal, all of which is during continuous data acquisition.
    I am able to actuate the switch only when pressing the "run" button in the DAQ assistant window. I am proficient with data acquisition but have never tried programming an output... Please help!
    Thanks!
    -a troubled researcher
    P.S. I am running LabVIEW 8.5 as well.

    blsmith4,
    You probably won't get the control of the digital port on your card by only using the DAQ Assistant. One of the following examples should provide you the functionality that you would like out of the box:
    Digital - SW - Timed Output (Simple)
    Digital - Continuous Output (More Complex)
    Continuous Write Digital Port - External Clock - Non Regeneration (Most Complex)
    Let me know if these work better for you.
    Best,
    Jason M.
    Applications Engineer
    National Instruments

  • I do not see the DAQ Assistant Express on Functions Input palette.

    Software config: Windows XP machine, Traditional NI-DAQ 7.4.4 for PCI 6602. I do not see the DAQ Assistant express on Functions>>Input palette. I get error message when I place the Instrument I/O assistant on the block diagram:-" Measurement and Automation Explorer or Instrument I/O Assistant not installed correctly. Please install from LabVIEW CD". I've tried uninstalling, re-installing driver and rebooting my computer several times.

    Hello PJ-SJM,
    The DAQ Assistant Express VI installs with the NI-DAQmx Driver and not Traditional NI-DAQ.  The good news is that the PCI-6602 is supported in both Traditional NI-DAQ and NI-DAQmx.  Check this KnowlegeBase article to find out what version of NI-DAQmx you need and then search the Drivers and Updates page for to find the version you need.  Once you download and install the driver you should see the DAQ Assistant Express VI on your functions pallet.
    The Instrument I/O Assistant is not for use with DAQ cards, it is for using other communication protocol like serial or GPIB.  However, if you need this you can install the Device Driver CD and it should work after that.  Note:  The device driver CD that came with LabVIEW should also contain the correct version of NI-DAQmx.
    I hope this helps, and have a great Friday!
    Brooks

  • Can't open properties window of DAQ assistant express VI.

    I have a program which uses DAQ assistant express VI's, reading several channels from a NI PCI-6220 card. Its been working fine for months but now some of the channels have started acting up, noise and oscillations that are not real. I've solved this problem by changing channels before but now when I try to open up the properties window of the DAQ assistant express VI to change channels, by double clicking the icon, or by right clicking and selecting properties, nothing happens! The cursor changes to a hourglass for a second and no window pops up. I've recently upgraded from 7.0 to 8.0 but that shouldn't make a difference, and even when I tried creating a new VI, and a new DAQ assistant express VI it won't work on this either. I've tried looking at security settings etc. but no joy. Any ideas?

    I would try uninstalling DAQ and reinstall.  Not scientific but in most cases will fix the issue of the DAQ assistant. 
    Matthew Fitzsimons
    Certified LabVIEW Architect
    LabVIEW 6.1 ... 2013, LVOOP, GOOP, TestStand, DAQ, and Vison

  • DAQ Assistant: Clock Settings (Samples To Read, Rate) can affect signal readings?

    Dear all,
    I'm totally new to Labview, and recently I get confused with the equipment I'm dealing with.
    Technical details:
    The Labview version is 7.1; computer operate system is Windows XP;
    The equipment has a NI PCI-6220 and a 68 Pin Connector Block to read signals from the equipment;
    There are 4 channels in DAQ Assistant (2 pressure reading, 2 temperature reading);
    For the first pressure reading, Signal Input Range from 4m to 20m Amps;
    Clock Settings are Samples To Read = 5, Rate (Hz) = 20.
    Description of the problem:
    I use Labvew to monitor and record pressure readings and temperature readings. The Labview configuration was set up by my advisor several years ago. Recently, I found the pressure reading vibrated a lot; for example, 5.01 to 5.05 bar within a second. In order to get a stable pressure reading, my advisor suggested me to change "Clock Settings" in DAQ Assistant from Samples To Read = 5, Rate (Hz) = 20, to Samples To Read = 250, Rate (Hz) = 1000. In this case, she believed that since we increase sample numbers and sampling rate, we could have more data, and thus have stable pressure readings.
    At first I could have very stable pressure reading. The last digit (0.01) did not change within 20 seconds. However, somehow after a day the pressure reading became unstable and even worse than previous. (pressure reading vibrates from 5.01 to 5.30 within a second)
    This is not the worst case. We found that when we set Clock Settings: Samples To Read = 5, Rate (Hz) = 20, the pressure reading is about 8 bar. However, when we set Clock Settings: Samples To Read = 250, Rate (Hz) = 1000, the pressure reading is about 5 bar. In this case, we even don't know which pressure reading is correct.
    Labview records current, and transforms it into pressure reading. Thus my advisor tried to monitor the current reading by Labview, and she found the current reading changed when she changed the Clock Settings. (0.004 Amps (5 bar) when Samples To Read = 5, Rate (Hz) = 20; 0.005 Amps (8 bar) when Samples To Read = 250, Rate (Hz) = 1000)
    Since we only change the sample numbers and sampling rate, the average readings should still be similar. However, the reading are not similar. That is what confuses me.
    My questions are, if Clock Settings in DAQ Assistant could affect signal readings? If so, how it could affect the signal readings? What is the effect of "Samples To Read" and "Rate (Hz)"? How to determine these parameters to get the true pressure readings?
    Thank you very much for your help. Hope to have some feedbacks from you.
    Best regards,
    Cheng-Yu
    Energy and Mineral Engineering
    the Pennsylvania State University

    A 6220 cannot read a current, it can only read a Voltage, so you'll probably have some (or should have) a resistor accros the voltage input. (normally 50 Ohm for a 0-20 mA signal).
    My first step would be to measure this voltage with a multi-meter so you know what the actual voltage should be.
    Then I would read that same voltage with MAX (measurement and automation explorer) to make sure you have the right value
    Now about the changing of the voltage/current/pressure, how have you terminated the other signals? Have you provided a good earthing?
    If you sample with a high frequency (1 kHz), perform an FFT on the acquired data, I can imagine a dominant 50 or 60 Hz (depends on where you live) in the signal that might cause your problem.
    Ton
    Free Code Capture Tool! Version 2.1.3 with comments, web-upload, back-save and snippets!
    Nederlandse LabVIEW user groep www.lvug.nl
    My LabVIEW Ideas
    LabVIEW, programming like it should be!

  • DAQ Assistant, Analog Signal to Analog Comparison Event for sample clock

    I am using a PCI-6071E and DAQmx.  I would like to use an analog signal on PFI0 to trigger Analog Comparison Event to be used as the sample clock in DAQ Assistant.  Is this possible and if so, how? 

    Hi Ben,
         You can use your analog signal on a PFI0 line as you are intending.  A few pitfalls of doing so are listed in this article.  Additionally, this Developer Zone article explains what you are trying to do, and gives and example that you may find helpful.  Have a good one!
    aNItaB
    Applications Engineer
    National Instruments
    Digital Multimeters

Maybe you are looking for