Using PCI-6251 with Labview 6.1

I recently purchased a PCI-6251 multifucntion DAQ board to get a higher sampling rate than my old PCI-MIO-16E-4 board.  However I have existing software that I want to run on the new board.  The board came with NI-DAQmx which will only work with Labview 7 or above.  I am currently running NI-DAQ 6.9.3 and the Measurement and Automation program does not recognize the new PCI-6251 board.  Is there anyway to get Labview 6.1/NI-DAQ 6.9.3 to recognize and configure the new board?  Or will I be forced to upgrade my Labview?

The M-series boards (PCI-62xx are M-series) require NI-DAQmx driver. The NI-DAQmx driver requires a newer Labview. No way around it.
John Weeks
WaveMetrics, Inc.
Phone (503) 620-3001
Fax (503) 620-6754
www.wavemetrics.com

Similar Messages

  • Is Ardence RTX required when I use PCI-7831R with labview real-time module for RTX target

    I am new in the labview real-time field. I want to use labview real-time module for RTX target on only one desktop. My card is PCI-78331R.  Is Ardence RTX necessary? Why I cannot find my card in the MAX after I configured the RTX?
    Thank you very much.

    Hi Ziaozhongda,
    You must have the RTX runtime engine installed before the LabVIEW Real-Time RTX module will install.  Additionally, you should follow the instructions in the "Using the NI PCI-7831R with the LabVIEW Real-Time Module for RTX" section of Getting Started with the NI 7831R (page 8).
    Please see Configuring Hardware for Use with RTX.
    Also useful is "How Can I Validate That My PC Will Run RTX?

  • Can I use a DAQ PCI-1200 with Labview 5.0.1?

    I have a Labview 5.0.1 software, and I want to use PCI-1200 with it. Can I do that? My OS is Windows XP, but I can change it with Windows98/2000.

    Hello jhylee,
    The latest NI-DAQ driver for the PCI-1200 is 6.9.3, which is compatible with LabVIEW 5.0.1. LabVIEW 5.0.1 supports Windows 2000, but not Windows XP (although it may still run).
    Have a nice day!
    Robert M
    Applications Engineer
    National Instruments
    Robert Mortensen
    Software Engineer
    National Instruments

  • Can I use IF-RIO with LabVIEW 8.20

    I am going to upgrade my LabVIEW to 8.20 but I am not sure whether the driver of IF-RIO can work with it properly.
    Does anybody know this?

    Hi,
           My customer get problems when using PCI-5640R in LabVIEW 8.2, but they could work in LabVIEW 8.0. Is there any latest NI-5640R driver support LabVIEW 8.2? Or PCI-5640R only works in LabVIEW 8.0? 
    Operation system and environment:
    LV 8.2
    LV FPGA 8.2
    NI-5640R 1.1
    NI-RIO 2.1.3
    Display error message in LV 8.2 FPGA complier server.
    An internal software error has occurred. Please contact National Instruments
    technical support at ni.com/support with the following information:
    Error 1003 occurred at Open VI Reference in
    nirviGetModInfo___nirviEIOImplementation_modGen.vi​->nirviGetModInfo___nirviE
    IOImplementation_modGen.vi.ProxyCaller
    Possible reason(s):
    LabVIEW:  The VI is not executable. Most likely the VI is broken or one of
    its subVIs cannot be located.  Open the VI in LabVIEW using File>>Open and
    verify that it is runnable.
    VI Path: C:\Program Files\National Instruments\LabVIEW
    8.2\resource\RVI\StockIO\private\StockFPGA_io_modG​en\nirviGetModInfo_StockFP
    GA_io_modGen.vi
    Please refer the attached reproducable example code.
    Best Regards
    NITW AE
    Gary
    Attachments:
    UsingFilterCore.vi ‏550 KB
    filt20070509104003.zip ‏255 KB

  • Controlling PCI-1422 with Labview instead of MAX

    Hello Experts,
    I want your help in one of my problem. I have a camera from Sensors Unlimited which is SU-640 1.7 RT-D and a NI parallel card PCI-1422. I can use and acquire data with MAX from the camera but I want to operate,use and acquire images from the camera using LabView. Is there any way of exporting or using the camera with Labview instead of writing a fresh code in LabView? 
    Does LabView offer any support or reads automatically from PCI-1422 or any possible solution that I get the block diagram code for using this camera in LabView.
    Please let me know, reply.
    Any help or suggestions will be highly appreciated.
    Thanks in advance!
     

    Nitai wrote:
    Thanks but again it requires writing and struggling with coding and experience on  LabView skills.
    Just wondering something easy. I tried one of the example and it worked actually but I am trying to get more ideas or options to make it look better and impressive.
    You want something and easy and doesn't require any effort on your end?    Hire someone who knows how to program in LabVIEW.

  • Is anybody use Hyper File with LabView?

    Is anybody use Hyper File with LabView?
    I Have a project that need to conect to a hyper file data base Made by PCsoft.
    Thanks
    Benoit Séguin
    Software Designer

    I have never used HyperFile, but a quick check of the company's website indicated that there are ODBC drivers available for it so it should be a simple matter connecting through ADO.
    Mike...
    Certified Professional Instructor
    Certified LabVIEW Architect
    LabVIEW Champion
    "... after all, He's not a tame lion..."
    Be thinking ahead and mark your dance card for NI Week 2015 now: TS 6139 - Object Oriented First Steps

  • How to generate a pulse of 5 V -ve polarity in synchronis​ation with a 10 v input using PCI 6251 DAQ board

    HI
    I want to generate a 5V -ve polarity pulse to trigger my IEEE 1394 camera using an PCI 6251 (scb68 pin E) device, furthermore I am reading a standard Vsync signal from a SVGA port which is 10 v.
    Now I want to synchronise the image capture of an IEEE 1394 camera with the Vsync signal so that camera is triggered to capture at every 4th Vsync signal (which is 50HZ.). Presently I am able to read the Vsync signal and also generate the -5 v signal But the Daq board generates a continues -5v signal while I want a 1 ms pulse (for every 4th Vsync).
    Also this 5 v signal is generated even if the vi is stopped running.
    Can anybody guide me so this problem can be solved.
    Thanks a lot in advance
    Shri

    You can use the Write function, that is you can use the 10 volt signal as a trigger for the task, and when you get this trigger, the task starts and generate (writes) a waveform you defined in an array ("data", in the example).
    Something like this..
    DAQmxCreateTask("",&taskHandle);
    DAQmxCreateAOVoltageChan (taskHandle, clokSource, your_sampling_freq, -10.0, 10.0,
    DAQmx_Val_Volts, "");
    DAQmxCfgSampClkTiming (taskHandle, clockSource, SAMPLING_RATE,
    DAQmx_Val_Rising, DAQmx_Val_FiniteSamps,
    samp_to_read);
    DAQmxCfgAnlgEdgeStartTrig (taskHandle, input_channel, DAQmx_Val_RisingSlope, 9.0);
    DAQmxWriteAnalogF64 (taskHandle, 1000, 0, 10.0,
    DAQmx_Val_GroupByChannel, data, //data contain a pulse
    &samp_per_channel, 0);
    DAQmxStartTask(taskHandle);
    // and then you create a loop....
    For example you can specify data to be like this:
    data[0] = -5;
    for(i=1;i<1000; i++){
    data[i] = 0;
    Tell me if you need more help.....
    bye

  • Closed loop control of DC motor drive using PCI 6251

    Hi
    i am using NI PCI 6251 for closed loop control of dc drive.(shunt motor,220v, 0.5HP, 2.5A rated, 1500RPM)
    the dc motor is fed from a bidirectional dc-dc converter. this converter works in boost mode(forword motoring) and buck mode(braking mode/battery charging)
    the converter is built using mitsubishi IGBT module. at low voltage side of the converter a battery bank of 48V is connected and high voltage side the motor is connected.
    An IR sensor is developed to measure the speed of the motor and a hall current sensor is used to sense the converter current.
    in the control part i have an inner current control loop(PWM controller) and a outer speed control loop.both controllers are PI controller. To measure the speed i have used counter of the DAQ assistant, then frequency. then this frequency is converted to RPM using labview.after giving the speed and current signals  to the controllers , then i started  tunning the PI gain, but at the starting the duty cycles was 100%, so the IGBT got damaged immediately. now how to tune the controllers during the closed loop operation.  the switching frequency of the converter is 20KHz.   
    can any body suggest me how to run the dc drive in closed loop ?
    thanking you.

    Hi Premenanda
    The freq data receiverd from the daq assistant has the units of Hz, or in terms of motor speed, rps (rotations per second). I can see that you are trying to convert that to rpm (rotation per minute). which can be done by dividing the rps speed by 60. Instead, you are multiplying it by 60 and also comparing with 2000. the resultant multiplied value is displayed in your program only if it is smaller than 2000.which is causing the problem here.( as seen in file 'original code.png)
    I have attached a file named 'modified code.png' that shows how the conversion from rps to rpm is done. Please do the required changes in your code and let me know if it works.
    Attachments:
    modified code.png ‏18 KB
    Original code.png ‏15 KB

  • Synchronous UART using PCI-6251

    Hello
    I am a beginner using LabVIEW and I need to create a synchronous UART using the PCI-6251, meaning, using one digital I/O to generate clock while another one shifts the data bit-by-bit. It has to work as a master in a half duplex mode. (LabVIEW always providing the clock but data flowing in one direction at a time).
    Could someone help me?
    Thanks!

    Hello, 
    It looks like your application is similar to the one in this forum. However the hardware in the forum is an X-Series in which the digital I/O has a dedicated sample clock and the PCI-6251 is an M-Series which requires you to provide the sample clock.  In order to accomplish what you are describing you will need to create a counter output task to use as your sample clock for the digital output and also to export on a PFI line.  An example of using a counter to correlate the digital output can be found by opening LabVIEW and navigating to Help»Find Examples»Hardware Input and Output»DAQmx»Digital Generation and selecting the Correlated Dig Write With Counter.vi.
    Thank you,
    Justin P 
    Justin
    National Instruments
    Product Support Engineer - Conditioned Measurements

  • PCI-6251 into LABVIEW DAQ Assist block interrupti​ons and resets

    Signals being received from PCI-6251 into a LABVIEW DAQ Assist block interrupts and resets every 10 seconds after 1 minute of data acquisition. Would this be a memory buffer problem or is it something related to the hardware? I am using a BNC-2110 hardware connector. Please view the attached image.
    Attachments:
    Interuption1a.GIF ‏4 KB

    Hey Peter,
    Ahh, I understand the graph now!  Hmm, this is very strange behavior.  Instead of using the DAQ Assistant, try to use the explicit DAQmx VIs (see picture).  You have more explicit control over what is going on with the DAQmx VIs and keep as much as you can out of the loop.
    Regards,
    Erik
    Attachments:
    analoginput.JPG ‏23 KB

  • Using Agilent 82357a with Labview

    Is it possible to use Agilent 82357a USB-GPIB interface board with Labview
    Linux version 7 ?

    Hi
    I just installed Labview 7.1 for linux on my machine after throwing away (not literally !) the Agilent card and borrowing an NI PCI-GPIB card. This works with kernels 2.4.x (tested with /usr/local/vxipnp/linux/NIvisa/NIvisaic). However, for proper operation of my machine (have bad display and mouse problems with the old kernel)  I need kernels 2.6.x. The driver in /usr/local/nigpib complains that it cannot work with those. I tried hacking the INSTALL script but it keeps running up against differences (like the modversions.h file) between the two kernel series.
    Can I use a newer version of the gpib libraries for linux with an older version of Labview like I am using ? I guess I will also need to get newer versions of the visa drivers for that. Though, the labview installation went smoothly with kernel 2.4.27-3 on my Debian system, I know that the newer versions of nigpib (is that the same as ni48825L.iso ?) and NIVISA are not supported on Debian. I cannot reinstall with a different distro, so please do not suggest that.
    Any other suggestions ?
    Thanks.

  • Integrating custom PCIe hardware with LabView

    Hi, I am new to LabView so appologize if question is in the wrong place.
    I have a custom PCIe acquisition card that I want to use in LabView environment to postprocess some data. So far the board has been used successfully with its windows drivers and VS.NET software. What is the process for setting up a device like that in LabView? Can LabView use the COM object windows driver just like a Visual Studio project would? If yes, is there a document that explains how this is done?
    I have worked with NI's own instruments where the driver is provided and seamlessly integrates into the LabView, but I am a bit lost with this custom non NI board.
    Thank you.

    Thanks. I tried searching that site for examples... but this is a board that our company has developed (software defined radio) and I am certain no one has written any drivers for it yet. Thats the problem statement that I am trying to research. What would it take to make the board we designed compatible with LabView. Since this is a Software defined radio platform, LabView offers some useful functions that can be used to postprocess and analyze data captured by our card.
    Ideal goal is to have LabView module capture data from PCIe buffer allowing for further processing (filtering, FFT, demodulation, etc...). I would appreciate if you can direct me to some examples or literature on this topic.
    Thanks,
    Anton

  • The Problem about Monitoring Motion using PCI-7358 on LabVIEW Real Time Module

    Hello everyone,
    I have a problem about monitoring the position of an axis. First let me give some details about the motion controller system I’m using.
    I’m using PCI-7358 as controller and MID-7654 as servo driver, and I’m controlling a Maxon DC Brushed motor. I want to check the dynamic performance of the actuator system in a real time environment so I created a LabVIEW function and implemented it on LabVIEW Real Time module.
    My function loads a target position (Load Target Position.vi) and starts the motion. (Start.vi) then in a timed loop I read the instantaneous position using Read Position.vi. When I graphed the data taken from the Read Position.vi, I saw that same values are taken for 5 sequential loops. I checked the total time required by Read Position.vi to complete its task and it’s 0.1ms. I arranged the loop that acquires the data as to complete its one run in 1ms. But the data shows that 5 sequential loops read the same data?

    Read Position.flx can execute much faster than 5 ms but as it reads a register that is updated every 5 ms on the board, it reads the same value multiple times.
    To get around this problem there are two methods:
    Buffered High-Speed-Capturing (HSC)
    With buffered HSC the board stores a position value in it's onboard buffer each time a trigger occurrs on the axis' trigger input. HSC allows a trigger rate of about 2 kHz. That means, you can store a position value every 500 µs. Please refer to the HSC examples. You may have to look into the buffered breakpoint examples to learn how to use a buffer, as there doesn't seem to be a buffered HSC example available. Please note that you need an external trigger-signal (e. g. from a counter of a DAQ board). Please note that the amount of position data that you can acquire in a single shot is limited to about 16.000 values.
    Buffered position measurement with additional plugin-board
    If you don't have a device that allows you to generate a repetitive trigger signal as required in method 1.), you will have to use an additional board, e. g. a PCI-6601. This board provides four counter/timers. You could either use this board to generate the trigger signal or you could use it to do the position capturing itself. A PCI-6601 (or an M-Series board) provides can run a buffered position acquisition with a rate of several hundred kHz and with virtually no limitation to the amount of data to be stored. You could even route the encoder signals from your 7350 to the PCI-6601 by using an internal RTSI cable (no external wiring required).
    I hope this helps,
    Jochen Klier
    National Instruments

  • Using PCI-8532 with NI-DNET 1.6.6 / NI-DNET 2.0.2

    I have a test fixture PC with the following configuration:
    PCI-8532 DeviceNet card
    Windows 7 Enterprise, SP 1
    LabVIEW 2010 Runtime
    NI-DNET 1.6.6 with MAX 5.0
    No LabVIEW development environment
    PCI-8532  is not available in MAX and the Windows Device Manager shows the card as 'PCI Device' with exclamation mark.
    Here is a screen shot of the Windows Device Manager and MAX:
    I cannot install the NI-Industrial Communication for DeviceNet 2.0.2 on this computer because I am getting the following error:
    Well, I have two questions
    How can I install and use the PCI-8532 and see it with NI-DNET 1.6.6? If this is not possible:
    How can I create an executable on my development system (laptop) with LabVIEW 2010 or LabVIEW 2011 and start this executable on the test fixture. Currently on my development system I have:
    LabVIEW 2010 and LabVIEW 2011 professional
    NI-DNET 1.6.6
    No NI hardware at all
    Do I need to install the NI-Industrial Communications for DeviceNet 2.0.2 on my laptop (my development system)
    Will this interfere with my current NI-DNET 1.6.6 and can I select which driver to use at compilation time
    Thanks,
    Nick
    Solved!
    Go to Solution.

    Hy Nick,
    No, it is defenitely not a replacement yet and I will work with our Web Department to make that more clear on our web page.
    The development for the new API is pretty much done but based on your feedback we are thinking of improving the compatibility so you could for example use the old 1.6.x APi and the new 8532 APi on the same machine, so you could use old and new hardware on that same machine. Would that be important for you?
    For now we are planning to allow only one hardware type at the same machine to save some development time. Let me know what you think.
    To clarify the today situation: The InCOM for Devicenet  component is not a part of LabVIEW. It is just a very simple way of communication with IO variables and function blocks for EM. The driver should install just fine even without LV installed. The error message is pretty much a Bug on our side and the workaround would be to use the LV Installer builder to create a new Installer that can install the Incom Dnet 2.0.x driver without having Lv installed.
    And today you should be able to have both the 1.6.x and the 2.0.x driver installed in parallel and use the 1.6.x API with ypour old boards and the 2.0.x IO variable approach with your new board.
    I keep this post updated as soon as we have a stable Beta available I will post something.
    DirkW

  • Using a driver with labview

    I want to create a GUI using labview to control a data acquisition card using a driver that was not created in labview. Could someone point me in the direction of some documentation that would give me some info on doing this.
    Thanks

    If you're using a non NI data acquisition card and the manufacturer doesn't have a LabVIEW driver, then you'll probably be using either a DLL or ActiveX interface that they provide. For interfacing to third praty DLLs, start with "Using External code in LabVIEW" that is part of the LabVIEW Bookshelf. You will need details from the manufacturer on the function calls. There are some shipping examples on ActiveX. The Invoke Node and Property Node functions on the Communication>ActiveX palette are used to access methods and properties.
    If you were to mention the make and model of the board you're using, you might find someone here with knowledge of it that would provide more details.

Maybe you are looking for