DAQ Assistant, Analog Signal to Analog Comparison Event for sample clock

I am using a PCI-6071E and DAQmx.  I would like to use an analog signal on PFI0 to trigger Analog Comparison Event to be used as the sample clock in DAQ Assistant.  Is this possible and if so, how? 

Hi Ben,
     You can use your analog signal on a PFI0 line as you are intending.  A few pitfalls of doing so are listed in this article.  Additionally, this Developer Zone article explains what you are trying to do, and gives and example that you may find helpful.  Have a good one!
aNItaB
Applications Engineer
National Instruments
Digital Multimeters

Similar Messages

  • How to use DAQ Assistant to display two analog channels in a XY Graph

    Hi,
    I am new to LabVIEW. I have been assigned a project in which i have to display two analog signals from an analog potentiostat i.e. voltage and current (also a voltage signal) on a graph using labview. I am using a myDAQ device to acquire these analog signals and I am using labview 2013 to display these signals on a graph. I am using DAQ assistant to configure these signals. Rightnow I have separate graphs for each channel. However, I want a single XY graph to display one voltage as a function of the other i.e. I want to have voltage on the X-axis and current output on the Y axis. Any kind of help is greatly appreciated.
    I have attached a screenshot of the VI. that I have made.
    Regards,
    Max
    Attachments:
    New Microsoft Word Document (2).docx ‏32 KB

    Thanks for the quick response. I was able to make a X-Y graph. However, I can only see a dot floating around. It's not drawing anything. I want to see it making a plot as it collects data. Also, I want to see the complete waveform formed when I stop the VI.
    Thanks,
    Max

  • Link digital and analog signal?

    Hi, I am trying to "link" a couple of analog signals to a dital signal.
    I have a torque transducer with a integral quadrature encoder.  I am also monitoring a couple of other analog signals.
    The 3 analog signals are being monitored through simultaneou sampling: 1Mhz for ~1sec
    In these analog signals i am capturing certain events (peaks, etc).
    What I would like to do is tie these events to the encoder which is tracking "angle" (rotational position).
    For example, I determine that the peak torque occured at index 579,397 of the 1,000,000 element array.  What was the angle of the encoder at index 579,397 ????
    I am using 7.1
    Thanks for the help!
    Ryan
    Ryan
    LV 7.1

    Ryan,
    Because we share the same name, I've created an example for you.
    Basically,  we're using the analog input sample clock to latch the
    position of the encoder (connected to the counter). In a finite
    acquisition, the exact number of samples you request are generated on
    the sample clock. Since we are starting the counter task first, we'll
    be waiting on the analog input sample clock to start. This way, the two
    tasks will be exactly synchronized. Please let me know if you have any
    additional questions.
    Good luck!
    Ryan Verret
    Product Marketing Engineer
    Signal Generators
    National Instruments
    Attachments:
    AI_Encoder_Sync.vi ‏99 KB

  • Analog Signal to Digital a Primer

    Many users have analog material sitting around, and they would like to get it into digital form, so that it can be edited. As this material is below SD (Standard Definition) quality, I will address the methods for digitizing it, so that it can be best edited in a DV Project. The output can be a DVD-Video, or other format, depending on what one wishes.
    The process of getting this analog material into a digital form, is referred to as Capturing, and it is digitizing the analog material and writing it to a file, that can be edited.
    One will need some method of getting the analog signals into digital form, and for connecting the analog output from their analog playback device to the computer. Though the analog signal could come from all sorts of devices, I’ll use a VHS tape deck in this example.
    Since a VHS tape deck does not have an A-D (Analog to Digital) processor, and also does not have a a digital signal Out, one needs something in between it, and the computer. That “something” serves two needs: mechanical connection of cables and the ability to convert the analog to digital for use in an NLE (Non Linear Editor). That “something” can be a digital video camera with pass-through capabilities, a device, called a “bridge,” or a capture card inserted into a bus slot in the computer. All three have pluses and minuses.
    The first two methods rely on one having a FireWire connection on the computer and will use the Capture module in the NLE software.
    The digital video camera route is good, as one already owns it. Having one with pass-through capabilities is getting more difficult. The camera also has to have FireWire (FW, IEEE-1394a, or iLink) output, and the computer must have a matching connection.
    An A-D bridge will do the same thing: allow one to connect the analog cables to one end/side, process the analog signals to digital and then output via FW to the computer. There are various such bridges available and two of the most often recommended makes/models are: the ADS-Pryo, and theCanopus/Grass Valley ADVC-110. Products, like the Dazzle lineup, just do not seem to work, or work well. I would avoid such.
    A dedicated capture card performs the same task, with one exception. The card will mount into an empty PCI slot on the MoBo, and have inputs for the analog cables. The card will have chips on it, to digitize the analog signals and feed these to the computer via the PCI slot. These do not connect via FW, but have one drawback: they must, in virtually every instance, use additional software to do the actual Capture, as most NLE’s cannot directly see and work with them. Some of the included software Capture programs are better, than others. One common issue with many of these cards is a loss of sync, between the Audio and the Video. This is usually static OOS (Out Of Sync), in that it is usually off by a fixed number of frames, and does not drift over time. This can be fixed in the NLE, but is extra work. This ARTICLE will give you tips on correcting OOS.
    Now, we are ready to begin the Capture, as our equipment and cables are all ready to go. We have a big decision in front of us - what format/CODEC do we use to Capture our digital files? Since we are doing an SD Project with non-HD (High Def) material, there is one fine choice: DV-AVI Type II w/ 48KHz 16-bit PCM/WAV Audio, using the DV/DVC CODEC. These files will be approximately 13GB / hour of Duration. Storage will need to be considered. These files are also I-frame, meaning that each “Frame” is rendered and available, so that frame-accurate editing is possible. The NLE will not need to do any processing, or conversion of these files, as even the Audio is ready for DVD-Video. Note: there will still be Conforming and the generation of PEK (Waveform Display) files, but the NLE will be able to do this quickly and efficiently.
    Using AVI Uncompressed will get you nothing that you can see, other than larger files. Do NOT use any compressed format/CODEC, like MPEG-2, WMV, MOV, etc., as these will loose data from an already inferior source, and will then need conversion to be used, either with a stand-alone conversion program, or internally by the NLE. You loose quality and then have more work to be done. Do not let the temptation of smaller files lure you into using anything but DV-AVI Type II, and just accommodate the larger file sizes.
    With the file format/CODEC decision made, the cables and devices hooked up, we’re ready to start the Capture. The exact operation will depend on what our equipment choice has been. See above for talk about additional Capture software, if one uses a capture card. Otherwise, we should be able to use the Capture module of our NLE. The first issue is that we will NOT have Device Control, as we would with a Capture from a miniDV tape-based digital camera, where we can control the mechanical aspects of the camera, during this process. As our VHS deck is hooked up to the A-D bridge, or digital camera w/ pass-through, via analog cables, our control signals cannot be sent to it. One would cue up the tape in the VHS deck, and then choose Record/Capture in the NLE’s Capture module, starting the Capture. Instantly, they would hit Play on the VHS deck, or via that deck’s remote control. This is easier the second time that one does it. I sit with the mouse in my right hand, poised over the Record/Capture button, and the VHS deck’s remote in my left hand, finger hovering over Play. It’s then a 1 - 2 push, and all starts. One can stop the Capture, if an unwanted section of tape comes up. The Capture module of the NLE will direct us on how to proceed from there, should more Captures be wanted. Some will just start up at a later point, and write an additional file for that portion of the tape, but some might require that one basically sets up a new Capture session One should investigate just what their NLE will require for doing multiple Capture sessions from one VHS tape.
    When done, one closes the Capture module, and then navigates to the location that these DV-AVI Type II files have been stored in, and Imports them into the Project. This differs slightly from doing the Capture from a miniDV tape camera, as those Captured files are already in the Project.
    Edit away, and Export/Share to the necessary output settings, file type and CODEC.
    What about those VHS - DVD decks, the ones that will burn a DVD directly from the VHS tape? They seem so simple to use, and for the most part are. However, those DVD-Video discs, contain pretty heavily-compressed MPEG-2 AV streams. These MPEG-2's, besides being compressed, are in aGOP structure and are muxed (Multiplexed into one file with both a Video & Audio stream), plus there is often some Menus and navigation in the first VOB, the main container components of a DVD-Video. Many NLE’s can Import the VOB’s, and basically rip the MPEG-2 out of them, but that first VOB can be highly problematic, due to the way that most of those decks do the Menus and navigation. It is seldom 100% DVD-compliant by the DVD-specs. This can be a major problem. This ARTICLE will give you some background and tips on DVD VOB's.
    Those dual-decks are good for doing one thing, regarding the DVD’s that they produce - provide one with an archival DVD-Video of the VHS tape, and not to be edited. However, the VHS side of these works well with an A-D bridge, just like a pure VHS deck does. I would strongly recommend that one does NOT use the DVD-Video portion of these, if editing will be desired. They cause many more problems, than they solve - just use them for an “archive” copy of the tape to a DVD-Video, for playback on a computer via DVD software or a set-top player, hooked to a TV.
    What about film? Well, film is obviously analog, and one needs to digitize this to edit on the computer. The process of this digitizing is referred to as telecineing the film. While one can buy equipment to do this, it is expensive, and involves a lot of setup and testing to get the job done. There are many service bureaus, that offer telecining. Some are better than others. The main considerations are: care in handling of fragile film, how the telecining is done, i.e. frame by frame w/ a liquid gate, or just by straight projection, and then output format for the digitized files. Some of these services will offer a DVD-Video disc. Do not be tempted to do this. See above. Others will offer to write these digitized files to a miniDV tape, and this is just fine. One does then need a miniDV camera, or miniDV tape deck, to then do the Capture from, but the quality and editability will be preserved. The final delivery format would be to put these DV-AVI Type II files onto an external HDD. This saves one from having to do the Capture from miniDV tape, having to have a miniDV camera, or miniDV deck, and only Importing of those files would be required to get them into the Project. Many of these service bureaus will either sell you the external, or use one provided to them. Check out specifications for things like format of the external. FAT-32 will limit the file sizes to ~ 4GB max, and NTSF will have no such limitations. However, if the service bureau is on Mac, and you’re on the PC, their Mac’s can ONLY see FAT-32 HDD’s. Talk this over with the service bureau first, so that there are no surprises later on.
    Hope that this helps,
    Hunt
    PS - also see this FAQ Entry for more detail.

    the_wine_snob wrote:
     Products, like the Dazzle lineup, just do not seem to work, or work well. I would avoid such.
    Although I now have a Canopus ADVC300 I previously used a Magix 'Movies on DVD' USB device. I found the results from this to be particularly good and, to my ageing eyes, I see very little quality difference.
    I think this is because Magix use their proprietary high quality codec (from their Pro products) to capture the footage. Like DV-AVI it runs around 10GB+ per hour. This can then, very simply, be exported to DV-AVI Type 2 for use in PRE. Note that there are some sound synchronisation issues with the 'out of the box' settings - but the fix is quick and I posted details to the Magix forum.
    Would I revert if I could return my ADVC300 bridge? Probably not - the firewire connection and the PC seeing the box as a camcorder adds to ease of use. In fact I am pleased to have both. There are some things Magix does better than PRE7 (batch exporting each scene as an individual DV-AVI file being just one) and it has smart-rendering which is handy for quick clip splitting/exporting.
    Cheers,
    Neale
    Insanity is hereditary, you get it from your children

  • DAQ Assistant: Clock Settings (Samples To Read, Rate) can affect signal readings?

    Dear all,
    I'm totally new to Labview, and recently I get confused with the equipment I'm dealing with.
    Technical details:
    The Labview version is 7.1; computer operate system is Windows XP;
    The equipment has a NI PCI-6220 and a 68 Pin Connector Block to read signals from the equipment;
    There are 4 channels in DAQ Assistant (2 pressure reading, 2 temperature reading);
    For the first pressure reading, Signal Input Range from 4m to 20m Amps;
    Clock Settings are Samples To Read = 5, Rate (Hz) = 20.
    Description of the problem:
    I use Labvew to monitor and record pressure readings and temperature readings. The Labview configuration was set up by my advisor several years ago. Recently, I found the pressure reading vibrated a lot; for example, 5.01 to 5.05 bar within a second. In order to get a stable pressure reading, my advisor suggested me to change "Clock Settings" in DAQ Assistant from Samples To Read = 5, Rate (Hz) = 20, to Samples To Read = 250, Rate (Hz) = 1000. In this case, she believed that since we increase sample numbers and sampling rate, we could have more data, and thus have stable pressure readings.
    At first I could have very stable pressure reading. The last digit (0.01) did not change within 20 seconds. However, somehow after a day the pressure reading became unstable and even worse than previous. (pressure reading vibrates from 5.01 to 5.30 within a second)
    This is not the worst case. We found that when we set Clock Settings: Samples To Read = 5, Rate (Hz) = 20, the pressure reading is about 8 bar. However, when we set Clock Settings: Samples To Read = 250, Rate (Hz) = 1000, the pressure reading is about 5 bar. In this case, we even don't know which pressure reading is correct.
    Labview records current, and transforms it into pressure reading. Thus my advisor tried to monitor the current reading by Labview, and she found the current reading changed when she changed the Clock Settings. (0.004 Amps (5 bar) when Samples To Read = 5, Rate (Hz) = 20; 0.005 Amps (8 bar) when Samples To Read = 250, Rate (Hz) = 1000)
    Since we only change the sample numbers and sampling rate, the average readings should still be similar. However, the reading are not similar. That is what confuses me.
    My questions are, if Clock Settings in DAQ Assistant could affect signal readings? If so, how it could affect the signal readings? What is the effect of "Samples To Read" and "Rate (Hz)"? How to determine these parameters to get the true pressure readings?
    Thank you very much for your help. Hope to have some feedbacks from you.
    Best regards,
    Cheng-Yu
    Energy and Mineral Engineering
    the Pennsylvania State University

    A 6220 cannot read a current, it can only read a Voltage, so you'll probably have some (or should have) a resistor accros the voltage input. (normally 50 Ohm for a 0-20 mA signal).
    My first step would be to measure this voltage with a multi-meter so you know what the actual voltage should be.
    Then I would read that same voltage with MAX (measurement and automation explorer) to make sure you have the right value
    Now about the changing of the voltage/current/pressure, how have you terminated the other signals? Have you provided a good earthing?
    If you sample with a high frequency (1 kHz), perform an FFT on the acquired data, I can imagine a dominant 50 or 60 Hz (depends on where you live) in the signal that might cause your problem.
    Ton
    Free Code Capture Tool! Version 2.1.3 with comments, web-upload, back-save and snippets!
    Nederlandse LabVIEW user groep www.lvug.nl
    My LabVIEW Ideas
    LabVIEW, programming like it should be!

  • Input of DAQ assistant does not initialize soon enough

    Dear all,
    We are trying to generate an analog voltage output with the NI 9263 and then read that as an analog input with the NI 9215 as part of an academic exercise. Both of these are on board the NI cDAQ-9174 (compact DAQ). An analog waveform is generated within Labview (either a sine, square, or constant), and that signal is fed into the DAQ assistant to make an analog output. 
    From there, wires connect the 9263 output into the 9215 input. A second DAQ assistant is used to acquire that input voltage from the 9215. 
    The problem occurs when the waveform generated voltage and voltage input are printed on a waveform graph. The input signal for the sine and square waves occurs before the waveform generation signal (i.e., the input is phase shifted to the left instead of to the right, as we would expect due to latency).
    We believe the problem occurs because the waveform is being output before the input has had a chance to initialize and start recording measurements. 
    Is there a way to initialize the DAQ assistant and start taking measurements before the output is generated? Or is there a way to determine the absolute time of each so that we can plot them in absolute time?
    Thank you for any help you can give.
    Attachments:
    Basic inout4square.vi ‏412 KB

    I tried cleaning up the program as best I could. I'm sure you can tell I am new with LabView. Thanks for taking a look even though it was so messy.
    The problem we are having now is that the input data lags the output data by anywhere from 30-50 ms, which seems like far too large a delay. This is when I turn the Waveform Graph to "Ignore Time Stamp". Because the generated output voltage is graphed from the waveform generator, it seems to take a while for it to progress through the program and through the hardware, thus the input voltage graph lags.
    When I turn the Waveform Graph to not "Ignore Time Stamp," the signal from the waveform generator occurs in the year 1903. I've tried searching on how to change this to the current time, but I can't seem to find a way to do that, despite plenty of other people wondering about this. Is there a way to do this so we can know the time at which the generated waveform signal is actually sent to the output DAQ assistant?
    Attachments:
    basic.vi ‏149 KB
    blockdiagram.PNG ‏28 KB
    delay.PNG ‏24 KB

  • Calibrating DAQ Assistant

    Let's say that I take a signal from my MPX4015A and measure it with analog input 3 on my DAQ board.  I go to DAQ assistant and find the previously-configured scaling for my MPX4015A to convert the voltage signal directly to kiloPascals.
    Now, how do I calibrate for a little bit of sensor inaccuracy?  I can go ahead and create tons of custom scalings, but it seems like there should be a little extra field that lets you calibrate the sensors using a certain measurement.

    Hello bmunden,
    I am unfamilar with the MPX4015A  sensor. I tried looking for a data sheet for this device and was unable to find any information on it when I did a Google search.
    What does this device do? Does it measure pressure and then output a small voltage? Does this device measure sound pressure or physical load pressure?
    DAQ assistant does have the something called the Virtual Channel Calibration, refer to the following link.
    NI-DAQmx Virtual Channel Calibration
    Please give this a try and let me know if this resolved your issue. Else please clarify and I'd be glad to assist you.
      Sandra T.
    Applications Engineer | National Instruments

  • Instr. and DAQ Assistants - Time to take off the training wheels?

    Hi everyone,
    I have written a control system program using a producer/consumer with state machine architecture - textbook LabVIEW beginner intermediate.
    In the producer, I have 6 different DAQ assistants taking AI and DI data from PXI mounted data acquisiton cards and two Instr. Assistants taking data from PXI serial ports.  In the consumer, I have a similar situation.  There are 7 DAQ assistants sending digital and analog requests to 3 different PXI chassis and one Instr. asssistant sending requests to a PXI serial port. 
    Is this frowned upon?  Would any of you pros call this a finished job if it is still bound with Assistants?  I have looked at the code involved in making these assistants into LabVIEW code (writing them out, so to speak), but it is complex and I haven't been able to put in the time to learn all of the different DAQmx and VISA functions and subVI's. 
    However, I have some difficulties that arise when, for instance, a component that is monitored/written to by a serial port is turned to manual control and the assistant times out.  It times out the entire control system, which is not desirable.  Is the fix in error handling, or getting rid of the assistants altogether?
    Any responses are appreciated - even if you are the bearer of bad news!
    Brad

    Hi Brad,
    Using DAQ assistants in a loop does come with some overhead. However if you are acquiring at relatively low speeds and do not want to use the DAQmx VIs, it shouldn't be a problem. Though programming using DAQmx VIs might look somewhat intimidating it follows a general priciple common to most I/O in LabVIEW. You open/create a channel, configure it, read or write and then close it. Generally, only the read or write VIs will be in the loop. So doing the creation and configuration every iteration of the loop is the overhead.
    I am not entirely sure what is timing out and what you are doing to cause a time out. There are two ways to generally fix time out errors. One, as you pointed out, is to handle the error. You can do this by unbundling the cluster to look at the error code, if it is a time out error, you can reset the error cluster.
    The other solution is to increase the timeout for the DAQ assistant. (Wire the time (in seconds) you want it to wait, to the timeout input of the DAQ Assistant).
    Cheers
    Malay Duggar
    NI

  • When using DAQ assistant to read frequency

    When using DAQ assistant to read frequency and Task timing is set to:
    N Samples, Clock settings to read 26,
    Frequency setup to rising edge,
    1 counter with 10 kHz to 1 kHz range.
    I get back a single number.
    Can I assume this is an average reading of 25 samples with the first sample unused?
    What is the base clock used?
    Is the “26” 26 cycles of the frequency to be measured?

    Hello,
    If you choose to acquire N samples from the DAQ Assistant then it will acquire all of these samples and return them to LabVIEW as an array.  However, when you use the DAQ Assistant it outputs the data in the dynamic data type first.  This data type makes it easy to graph and run the data through other express VIs.  If you were to create a Numeric Indicator from this data type it would just display the last element from the dynamic data array.  To display this data properly in a numeric format convert the dynamic data to an array of doubles by using the Convert From Dynamic Data function in LabVIEW.  Then you can select to convert it to a 1D array of scalars and when you create an indicator off of the output of this function all of the data should be displayed.
    The timebase that is used for lower frequency measurements is the onboard clock, which is internally connected to the Source.  Then you connect your signal to be measured to the Gate of the counter.  Since the frequency of the onboard clock is known it can be used to calculate the frequency of an unknown source based on when the counter is on and off (determined by the Gate). 
    Have a good day,
    Brian P.
    Applications Engineer

  • Error with 2 Daq assistant

    Hello
    Another noob joining the community. Here is my question.
    What am i doing wrong? I use 2 Daq assistants on the NI9401 module this is for digital input and output. I use it for generating 2 signals. If i test it with 1 it works but if i go to 2 or more it fails. This is the error i get . I have also added the program as an attachment.
    Thanks in advance.
    Solved!
    Go to Solution.

    Hello,
    That error means that the DIO task is reserved and cannot be obtained by one of the DAQ assistants. There is actually a nice example program on what you want to do. It can be found here:
    https://decibel.ni.com/content/docs/DOC-11632
    This example does not use the DAQ Assistasnt but the DAQmx API VIs (which the DAQ Assistant is using also anyway). If the order is not important (and since the these DIO tasks are software timed it is probably not an issue you can also choose to use the error cluster from one DAQ assistant and wire it to the other (see attached VI). Doing it that way you force one DAQ task to wait for the other one to finish. In that way they cannot run at the same time preventing for the error to happen.
    regards,
    Rik Prins, CLD
    Applications Engineering Specialist Northern Europe, National Instruments
    Please tip your answer providers with kudos.
    Any attached Code is provided As Is. It has not been tested or validated as a product, for use in a deployed application or system,
    or for use in hazardous environments. You assume all risks for use of the Code and use of the Code is subject
    to the Sample Code License Terms which can be found at: http://ni.com/samplecodelicense
    Attachments:
    2 Daq assistant generating error (3).vi ‏169 KB

  • DAQ-Assist crashes Labview....

    Hi,
      I'm running Labview 8.0.1
     I use the DAQ-Assist to read info from some channels.....
     I drag the icon on my block diagram, which automatically opens the wizard. I setup channels, type of signal (voltage, current, etc), hit OK and after ~30 seconds (where LV says first "Verifying task' and then "Building VI, please wait"), it creates an ExpressVI without any problems.
     The problem is whenever I try to edit/change/add settings.... Say that later I wanna change the voltage range on one of the channels, or add another channel. I go in (double click on the icon), make the changes and when I hit OK, LV crashes after about 15 or 20 seconds of displaying the "buidling VI, please wait" message.
     Usually I have to re-create the whole thing (drag a second icon, and delete the first one) to avoid the problem.
     I've seen this happen even with new VI's that I create from scratch. It happens 95% of the time, other times it works flawlessly.
     What's wrong?
    Attachments:
    DAQ-Assist-error.JPG ‏187 KB

    Hello Emerino,
    Thanks for contacting National Instruments. 
    It looks like there is an issue with the installation of either your DAQ driver.  The knowledgebase LabVIEW Application Error: The instruction at address x referenced memory at address y. The memory c... goes over possible causes of the error you are seeing.  Most likely a file in your LabVIEW or DAQmx install has become corrupted and is causing LabVIEW to crash. 
    First, attempt to repair all of your NI software.  This can be done by going to the Add or Remove Programs screen and clicking the change/remove button next to the National Instruments Software selection.  This will bring up a list of all of the installed NI software that is on your computer.  Highlight all of this software and choose to repair.  During this process the installer will ask for the LabVIEW CDs as well as locations of any installers that were downloaded to your hard drive. 
    If you are still having trouble running your DAQ assistant, update to the latest version of DAQmx.  This will have updates to numerous known issues as well as increase the stability of the DAQmx driver. 
    The Investigate Previous Internal Error window you are seeing after a crash is a way for you to report any unexpected errors in LabVIEW that caused it to shut down.  If you choose to report the error, National Instruments can use the information to improve its software products.  This window should only appear when opening LabVIEW after a crash occurs. 
    Please post back if you continue to have issues or if you have any questions. 
    Regards,
    Browning G
    FlexRIO R&D

  • Start button for DAQ Assistant doing pulse generation

    Hi,
    I am new to Labview, and I got the USB 6210.
    I used the DAQ assistant and selected ctr0 to create a pulse.
    This works fine, but the pulse is generated when you run the program.
    I want it so that when you press a button, the pulse generation starts. Then when you press the button again, it offs.
    How would you do this, I dont see any 'data' input anywhere on the DAQ Assistant. Please help.
    Also, I was wondering how to change the digital output line where the pulse will output. Currently it comes out the first D/O line. USB6210 has 4 digital outs, how do I change the output line.
    In DAQ assistant, all i see are two choices for pulse, crt0 and crt1 (USB 6210 has 2 counters).
    Again I am just learning, thanks for any replies

    Hi,
    I’m using NIUSB 6218 to acquire measurements from 16 (Ai16-Ai31) displacement sensors and 16 (Ai0-ai15) from ultrasonic sensors. These sensors are activated alternately which means when acquiring data from the ultrasonic sensors the displacement sensors are deactivated. For ultrasonic scan contains 16 channel x 16  samples  then it will followed by displacement sensors also 16 channels x 16 samples. The acquisition samples are controlled by the trigger signal (rising edge) from the PIC. The whole system is using external clock from the PIC 20MHz. I’m planning to insert START and STOP button on the front panel. The assignment of the pin is as follows:
                PFI0- Trigger signal (acquiring data) from PIC.
                PFI1- BURST signal from PIC (send all (16x16) acquired data to be displayed/save into spreadsheet file.
                PFI4- START signal to PIC
                PFI5- STOP signal to PIC
    So far my VI works when I applied external trigger signal at PFI0. The VI can sample the data but how to inform the PIC that the 16 sample data has been acquired? My problem is I do not know how to combine all these functions (START/STOP/BURST) into my VI. Attached is my sample of the VI which I’m working on. Please advice.
    TQ
    Attachments:
    displacement and ultrasonic data.jpg ‏259 KB

  • Acquiring pressure data using daq assist

    Hi all,
    I want to use DAQ Assist to measure the pressure.When i select DAQ Assist>Acquire Signals>Analogue Input>Pressure>Pressure(Bridge) then is says 'no supported devices found'
    What should i do?
    thanks

    I have Trafag pressure sensor anyway i give up doing it by daq assistant.I did it by using another daq tools.But now i have a new problem in calibration of sensor so i will post a new question

  • How to make a synchronous acquisition of two analog signals with a one channel DAQ ?

    Hi !
    It is the first time I use Labview. I have just made some easy VIs, and now, I do not know how to deal with my problem...
    My problem : I have only one acquisition card (DAQ Ni 6034E) and I would like
    to acquire simultaneously two analog signals. It seems to be possible, in a quasi-synchronous acquisition if the card acquires one point of the first signal, then the first point of the second signal, then the second of the first signal, and so on ... I thougth that I could made two sequencies with a VI of data acquisition in each sequency, with the AI MULT PT. But with this, I think that I will lose the precision of the sampling frequency. And I have to know the sampling fre
    quency ...
    Thanks for your help !
    Carline

    Hi Carline,
    A PCI-6034E is a low-cost board, which only has one Analog to Digital Converter (ADC).
    With only one ADC you can't acquire different channels at the same time. That's why you find a multiplexer before the ADC in this type of board. This enables you to acquire multi-channels at the same rate quasi simultaneously. The samples of the different channels will be interleaved as it is explained in the following knowledge base :
    - http://digital.ni.com/public.nsf/3efedde4322fef19862567740067f3cc/9379ea091c264b7c86256bc90082ca5d?OpenDocument.
    With this method, the sampling frequency is the same for all channels. The sampling frequency of one channel is determined with the "scan rate" parameter. There is only a small delay between the samples
    of each channels. When you perform an interval scanning acquisition, this delay depends on the "sample rate" that you specify. To have more informations about this terminology, please refer to the following KBs :
    - http://digital.ni.com/public.nsf/websearch/4D1435DF82EF494186256D8A006DD6D4?OpenDocument.
    - http://digital.ni.com/public.nsf/websearch/4D1435DF82EF494186256D8A006DD6D4?OpenDocument.
    To easily perform such a multi-channels analog acquisition, you can use the VI named "AI Acquire Waveforms.vi". You also can use an example provided with LabVIEW. You just have to browse the NI Example Finder in "Harware Input & output >> Traditional NI-DAQ >> General".
    Best regards,
    Benjamin
    National Instruments France

  • Synchronize Motion velocity with DAQ analog signal

    Hi Everybody
    I am acquiring analog signals using hw trigger and external clock with C and A signals of a 1024 counts TTL encoder.
    Now I want to measure the rotation speed of a shaft. I have a Flex-Motion board with which I want to read the velocity (or the same encoder), but I don't know how to synchronize each scan of the analog signals with the velocity signal for that instant.
    I can´t use RTSI bus because my Motion board doesn´t have RTSI connector.
    Software and hardware:
    LabView 6i (v. 6.0.1.b3)
    DAQ-Board: PCI-6023E, NI-DAQ 6.8.1, MAX 2.0
    Motion board: PCI-FlexMotion-6C, NI-FlexMotion 5.0.1
    Thanks a lot for your help.

    Hi chia,
    with the hardware configuration that you have, I think it´s not posible to synchronize both boards due to you cannot share the same clock source (not RTSI on the motion board).
    http://zone.ni.com/devzone/conceptd.nsf/2d17d611efb58b22862567a9006ffe76/66e6f4ee7e0848d686256a55006...
    Sorry for the delay :-)
    Regards
    DiegoM.

Maybe you are looking for