Do I have to use LabVIEW Real Time with a reflective memory node?

For reference with an external data system that will be temporarily installed at a customer's site, they have asked that I tie into their data network to record data from their control system.  They apparently use a reflective memory network for data sharing.  I have no prior experience with reflective memory, but all references to it involve real time systems.  I do not need absolute determinism to acquire this data, I can be late by several milliseconds with no problem.  Do I still need to use LabVIEW Real Time to interface with the PXI reflective memory node?

Hi AEI, 
I have worked with that card briefly before. It has a Visa based driver and RT isn't required. However, I haven't worked with the card on a non-rt system and am not sure if there any issues to be aware of. 
A lot of work has gone into integrating support for the card into Veristand, it may save you enough development time to use at an RT-Veristand system to be worth the extra cost. 
Jesse Dennis
Design Engineer
Erdos Miller

Similar Messages

  • Can I use Labview real time with PCI 6013?

    How much does Labview real time cost?
    Do i need to have some additional hardware to use labview real time with PCI 6013?

    You can not use the PCI-6013 board with LabVIEW Real-Time.
    We offer three platforms for LabVIEW Real-time. PXI embedded controllers, 7030 data-acquisition plug-in boards (7030/xxxx) and FieldPoint. For more information on these options go to www.ni.com/products and then choose Real-Time from the menu on the left. There are PCI versions of the 7030 boards. Each 7030 board has a data-acquistion board attached to it (6030E, 6040E or 6533). When you embed a LabVIEW Real-Time program on the 7030 you will be able to access the daughter board. You can not access other boards from the embedded program. If you want to use several daq cards in your real-time system I would recommend a PXI chassi and PXI daq boards.

  • Using JRockit Real Time with Large Heap

    I want to know if JRockit Real Time can work with extremely large heap, e.g. 100G, to provide effective deterministic GC, before proceeding with evaluation for my application.
    Thanks in advance!
    Wing

    Hi Wing,
    In general, extremely large heaps can make all GCs slower. This is true for Deterministic GC as well. However, it is very application dependent. It is above our standard recommendation, but you can download JRockit Real Time and give it a try.

  • Is Ardence RTX required when I use PCI-7831R with labview real-time module for RTX target

    I am new in the labview real-time field. I want to use labview real-time module for RTX target on only one desktop. My card is PCI-78331R.  Is Ardence RTX necessary? Why I cannot find my card in the MAX after I configured the RTX?
    Thank you very much.

    Hi Ziaozhongda,
    You must have the RTX runtime engine installed before the LabVIEW Real-Time RTX module will install.  Additionally, you should follow the instructions in the "Using the NI PCI-7831R with the LabVIEW Real-Time Module for RTX" section of Getting Started with the NI 7831R (page 8).
    Please see Configuring Hardware for Use with RTX.
    Also useful is "How Can I Validate That My PC Will Run RTX?

  • How do I talk to 1-Wire devices with LabVIEW Real-Time?

    Hello, I'm trying to use LabVIEW Real-Time to talk to some older 1-Wire devices that my company has previously communicated with using C code.  We are currently developing on a cRIO and are trying to use a Dallas Semiconductor 2480 serial port device to talk to some iButtons and a secure SHA 1-Wire device.  I have run into a few problems where the 1-Wire device stops communicating and just repeats the data sent to it.  I was wondering if anyone has successfully created any VI's to talk with 1-Wire devices via the serial port that do NOT use any kind of Java or .NET or Windows based device drivers supplied by Dallas Semi.  A last resort would be that we keep the C code and place it in a block in LabVIEW and run it as is but I would prefer to have some LabVIEW-native code.  Any help would be greatly appreciated.  Thanks.

    Brian_T,
        Thanks for that link about compiling a shared library for use on a VxWorks cRIO.  The only thing they mention that I don't understand is this bit:
    "GNU toolchain distributed with Wind River VxWorks"
    I have been unable to locate a compiler for VxWorks so I am assuming that is located in this GNU toolchain thing?  Does that come with the cRIO?  I think I noticed a VxWorks C compiler somewhere else online but we're talking a totally different product I would have to purchase, assuming it would even work.  I guess once the stub and the shared library are all good to go, I just call the C functions in the shared library and when it comes time to talk to the serial port, just slap a VISA Read/Write VI on the block diagram.  I've been using a cRIO and COM1 access for some time now so I do understand at least that much.  Let me know how to get ahold of this "toolchain" thing.  Thanks.

  • Desktop PC as LabVIEW Real-Time Target and DAQ ISA card

    I want to use LabVIEW Real-Time (Desktop PC + OS Pharlap ETS) for operation with DAQ ISA card.
    Pharlap ETS supports operation from Windows dll and can supports Third Party Device PCI.
    Whether there is a possibility to work with ISA card in Pharlap ETS?
    I want to work for example with card PC20TR from BMC Messsysteme GmbH.
    Does anyone have tricks for using ISA card in Pharlap ETS?

    Hi,
    If you have reviewed the requirements for a Real-Time target and verified that your computer is compliant you might want to check out this article.  The main issue that people encounter when using PC's as RT targets is have the correct network card.  Make sure the card you have is on the supported devices list on the webpage that Christian posted.  You should be able to develop a driver using low level VISA commands but it is not an easy task.  Read over the developer zone article to learn more about it.
    Eric A.
    National Instruments
    Distributed I/O Product Support Engineer

  • To develop an VI to aquire data of thermocouples from a remote chamber in labVIEW real time, should use NIDAQ or NIDAQmx?

    Hai
    I need to develop an VI to aquire data at real time. Am using LabVIEW RT as the OS and i need to know, if i can use NIDAQ 7.0 or should i use NIDAQmx? what is the diffrence between the two? is it just that NIDAQmx makes job simpler?
    Also can anybody please tell me, how to access a channel in labVIEW? am not clear about the path to follow to get the physical channel as a virtual channel on the LabVIEW.
    I am using a 8186 PXI with labVIEW RT and a SCXI, with both scxi and daq card on the same chassis.
    the development will be done on Windows XP which will be the host machine.
    Thanx in advance for any answers
    Arun

    Arun,
    DAQmx is our new Data Acquisition driver and should be considered as the primary option when writing new applications because of supportability in the future.
    DAQmx does make Data Acquisition tasks much simpler, abstracting many of the board implementation details and giving you a common API for 99% of the tasks you might want to accomplish.
    There are a couple of minimum requirements for using DAQmx on LabVIEW Real Time:
    - You will need LabVIEW Real-Time 7.1 (or higher)
    - You will need DAQmx 7.2 (or higher)
    These requirements are mentioned in the following Knowledgebase:
    Is NI-DAQmx Supported in LabVIEW Real-Time?
    The other thing you might want to make sure of is that DAQmx supports the DAQ hardware you already have or are planning to use. DAQmx does support most of the existing DAQ and SCXI hardware products, and reading thermocouples should become a very simple task.
    With DAQmx you can use physical channels, Global Channels (Aliases for preconfigured physical channels, equivalent to the Traditional DAQ Virtual Channels) or DAQmx Tasks (Preconfigured Data Acquisition tasks that include channel, timing and triggering information). A good place to start to learn about these features could be:
    Advantages of NI-DAQmx
    The following link might be useful to check Hardware support:
    http://www.ni.com/support/daq/versions.htm
    The driver can be downloaded by following these links:
    ni.com -> Support -> Drivers and Updates -> Current Software Versions -> Multifunction DAQ -> NI-DAQ Version 7.3 for Windows 2000/NT/XP
    This includes the latest version of DAQmx and our Traditional DAQ driver. Please note that DAQmx base is also listed but that's our Register-Level Programming (RLP) driver, which provides DAQ support for previously unsupported platforms.
    I hope this helps,
    Alejandro

  • How come you can not use more than 126GB of a 1.5TB diskdrive with LabVIEW Real Time?

    How come you can not use more than 126GB of a 1.5TB diskdrive with LabVIEW Real Time?
    This is only 8.3% of the available disk drive space!
    http://digital.ni.com/public.nsf/allkb/E7E2F5657B6​026358625750B007201A6?OpenDocument
    National Instrument's solution is either reformat you drive to less than 126GB or replace the drive with one less than that size.
    Please NI, tell me you are going to fix your LabVIEW Real Time soon, and that we won't have to wait for LabVIEW RT 9.2.1 in 2011 for this one to be fixed.
    http://upload.wikimedia.org/wikipedia/commons/9/90​/Hard_drive_capacity_over_time.svg
    Drives as large as 126GB or more have been around since about mid 2002.  That's about 7 Years now and LabVIEW Real Time STILL can't use drives larger that 126GB!  Please NI.  Do something.
    We deserve fixes to the old versions of LabVIEW RT.  How about updates to LabVIEW RT.  Such as 7.0.2, 7.1.2, 7.6.2, 8.0.2, 8.2.2, 8.6.2.
    We don't need new features or versions of LabVIEW RT such as 9.0 until we have some of bugs fixed in the existing versions. 
    Kevin. 
    Message Edited by kmcdevitt on 05-29-2009 03:39 PM
    Solved!
    Go to Solution.

    Coq rouge wrote:
    NI is not making the Real Time OS in house. I do not remember which firm that is making it. Anyway you are allowed to use a hard drive that has been partitioned. And 126GB is some space is it not? I remember Bill Gates declared one time that 640Kb ram should be more than enough for any PC application also in the future 
    Coq Rouge (formerly t06afre),
    The RealTime OS is an NI Product that is purchased from NI.
    NI may or may not outsource design, developement, or manufacturing of the products that it sells under its own Company name.
    For this reason I would be concerned if I were a PXI Embedded Controller designer/developer in the United States. 
    Yes, 126GB is some space.  I see that there are now 2TB drives available.  126GB would be 6.3% of a new 2TB drive.  Should we all be happy with being able to use 6.3% of the available drive space.  Moore's Law seams to be alive and well.  Drives larger than 126GB have been readily available for 5 or 6 Years.  This should not be a surprise to NI.  Why don't they do something about it.
    Thanks for the trivia on Bill Gates.  I must have missed that one since I don't try to keep up with Microsoft.  I believe that he is a wonderful person and he and his family are doing wonders for the world in their "retirement".   The world needs more like him.
    What I do remember is Steve Jobs telling us that we would never need more that 128MB and that we would never to need to do anymore software development because everything was already in the "ToolBox".
    McKevin.  

  • Can I use third party DAQ boards with Labview real time?

    I have an Adlink DAQ 2010 card that I curently use with labview. Will this card also work with Labview real time? What requironments are there for 3rd party cards to work with Labview real time?
    Thanks!

    Hello George,
         Here is a great link for you to take a look at: Can I Use a Third-Party Device with LabVIEW RT?
    Have a great day,
    NathanT

  • Could I use two regular computer to achieve realtime communicat​ion using the Ethercat and Labview Real Time Module?

    Could I use two regular computers (one acts master , another one acts as slave using the Ethercat Internet Card) to achieve realtime communication using the Ethercat and Labview Real Time Module? if it could, what hardware should I purchase from the NI ?
    Thank you!

    Hi Xiaolin,
    NI doesn't offer Windows based Ethercat master or slave software. Only LabVIEW RT can run the Ethercat driver. 
    However, you could use a LabVIEW RT target as an Ethercat master and use the Ethercat Internet Card with a slave PC (note: I think this will work from the Beckhoff description of the card you are describing. The card should integrate like any other non-NI slave. However I haven't tested the setup and don't advise saying it will work until you have tried). 
    You can use any NI RT target with two ethernet ports as the Ethercat master. This could be a cRIO, PXI or RT Desktop. 
    Jesse Dennis
    Design Engineer
    Erdos Miller

  • The Problem about Monitoring Motion using PCI-7358 on LabVIEW Real Time Module

    Hello everyone,
    I have a problem about monitoring the position of an axis. First let me give some details about the motion controller system I’m using.
    I’m using PCI-7358 as controller and MID-7654 as servo driver, and I’m controlling a Maxon DC Brushed motor. I want to check the dynamic performance of the actuator system in a real time environment so I created a LabVIEW function and implemented it on LabVIEW Real Time module.
    My function loads a target position (Load Target Position.vi) and starts the motion. (Start.vi) then in a timed loop I read the instantaneous position using Read Position.vi. When I graphed the data taken from the Read Position.vi, I saw that same values are taken for 5 sequential loops. I checked the total time required by Read Position.vi to complete its task and it’s 0.1ms. I arranged the loop that acquires the data as to complete its one run in 1ms. But the data shows that 5 sequential loops read the same data?

    Read Position.flx can execute much faster than 5 ms but as it reads a register that is updated every 5 ms on the board, it reads the same value multiple times.
    To get around this problem there are two methods:
    Buffered High-Speed-Capturing (HSC)
    With buffered HSC the board stores a position value in it's onboard buffer each time a trigger occurrs on the axis' trigger input. HSC allows a trigger rate of about 2 kHz. That means, you can store a position value every 500 µs. Please refer to the HSC examples. You may have to look into the buffered breakpoint examples to learn how to use a buffer, as there doesn't seem to be a buffered HSC example available. Please note that you need an external trigger-signal (e. g. from a counter of a DAQ board). Please note that the amount of position data that you can acquire in a single shot is limited to about 16.000 values.
    Buffered position measurement with additional plugin-board
    If you don't have a device that allows you to generate a repetitive trigger signal as required in method 1.), you will have to use an additional board, e. g. a PCI-6601. This board provides four counter/timers. You could either use this board to generate the trigger signal or you could use it to do the position capturing itself. A PCI-6601 (or an M-Series board) provides can run a buffered position acquisition with a rate of several hundred kHz and with virtually no limitation to the amount of data to be stored. You could even route the encoder signals from your 7350 to the PCI-6601 by using an internal RTSI cable (no external wiring required).
    I hope this helps,
    Jochen Klier
    National Instruments

  • Unable to install labview real-time

    I just successfully installed labview v8.0 on a pc and this is my first time to ever use this application. I was trying to install labview real-time v7.1 for my cRIO.  Unfortunately, it failed to install. I received an error message saying that I needed to have LabView 7.1 prior to the installation. LabView 8.0 is working great and I dont understand why does LV real-time wants an older version.
    Do I have to uninstall LabView v8.0 and order an older version to get labview v7.1 to get running?
    Any insights for a newbie will be greatful.
    Thanks 

    What do you mean by "I was trying to install
    labview real-time v7.1 for my cRIO"?  It is not clear if you got
    the error when installing that piece of software in the cRIO RT
    controller or when you were installing the software on your PC.
    Nonetheless, if your intention is to use LabVIEW 8.0, then there is no
    reason why you need to install anything that has to do with LV 7.1
    (such as cRIO RT support for LV 7.1 on your PC or LV RT -real-time- on
    the controller). In your case, it seems to be ok to ignore errors
    related to LV 7.1; it could have been that the computer you are
    installing NI software on had an older version of LV that was not
    properly uninstalled.
    BTW, the installers should be smart enough to install the right
    versions you will need given you install LV 8.0 first. To ensure all
    needed pieces are installed, make sure you run the installers in this
    order: LabVIEW first, followed by LV RT and/or LV FPGA, and NI-RIO last.
    Hope this helps.
    JMota
    National Instruments

  • Can I acquire and analyse in real time with regular Labview?

    I have to acquire samples (which vary cyclically in a roughly sinusoidal fashion) from a sensor, and check every sample to see if it is the minimum (the valley) of a cycle. If it is, and it does not fall within an expected range, I have to take a corrective action that involves rejecting the part that was just measured as well as a 30 more parts (to make sure that the defective part has been rejected). The signal from the sensor is not very noisy, but beause of the nature of the measured object, there could be local minima and maxima. To guard against that, a point is considered to be a valley only if subsequent readings deviate above that point by a certain amount. If a part is indeed defective,
    a digital out put has to issued to reject that part.
    Can all these be done using regular Labview (not RT)? I tried it out with a proto-type VI, using DAQmx vis, continously acquiring samples, but reading one sample at a time from within a loop (the VI I used is attached with this question). The result has been disappointing, since each time the loop executes there is a delay that keeps building up. Finally, even after the part feed has been turned off I can see Labview processing signals from parts that have long since gone past the measuring head.
    Another perplexing thing I found is that the time taken to execute the while loop in the vi is not consistent; it takes anything from 6 to 50ms to execute.
    I will need atleast 8-12 samples from a part to build its profile, and the feed rate is about 3000 parts per minute. I am using Labview 7.0 with an NI-6013 card in a Windows 2000 environment.
    Thanks for any suggestions / recommendations.
    Attachments:
    find_trough_2.vi ‏378 KB

    Hello,
    Thank you for your suggestions; I had already resigned myself to going for a Real Time system, your answer convinced me to commit myself to it!
    That said, your reply leads to a couple of (related) questions...
    1. Your point regarding the use of Local Variables is well taken; I have been repeatedly told at various training sessions how the necessity of updating the LVs during each loop iteration slows the computation time. However, what alternative do we have when there are several controls to which we have to write AND read data multiple times during a loop iteration, and perform different computations based on the value held by these controls? (You have seen the VI I attached with the original question). Some of these conditional
    computations further change the value of the controls. Does Labview have any other mechanism to store and manipulate the intermediate value from a computation?
    2. I did a simple experiment to determine the average loop time, and the results were surprising. I placed the entire content of the VI I used (Find the Valley in cycle.vi) in a stacked sequence structure, and wired the index counter "i" to a control to count the number of iterations the loop executes. I placed a frame before this with a tick count instruction to get the start time of the loop, and a frame after this to get the end time of the loop. Dividng the difference of these with the number of iterations, I got the average loop time to be around 1.2ms! Am I interpreting my results incorrectly?
    Thnaks once again for your response. I would really appreciate your views on the questions I have raised in this comment
    Regards
    Arun P. Madangarli

  • Read CPU temperature of Desktop PC running LabVIEW Real-Time

    Is it possilbe to read the CPU temperature of a Desktop PC running LabVIEW Real-Time? I found an older example that allows to read the temperature of PXI systems (http://www.ni.com/example/29538/en/). Is there a possibility for Desktop Real-Time Machines as well?
    Regards,
    Steffen

    Hi this is heavily dependent from your specific Hardware. Which PC or more specific which main board do you use?
    I found several examples in the NI forum that try to access the CPU data.  
    I’m using a Dell machine and both temperature VI’s are not working for me.
    Hope you have more luck.
    Attachments:
    Temperatur_der_CPU.vi ‏13 KB
    WMI_GetTemperatureProbe[1].vi ‏26 KB
    WMI_Win32procesor[1].vi ‏41 KB

  • End of memory. Labview + Real time module + Compact Field Point 2120

    My name is Michele Ciampichetti, I'm writing from Italy.
    Frist of all, excuse for my english.
    This is my first experience with Labview, and the application is not so easy. We use Labview 8.2.1
    Our system is costituited by two different parts: there is a Compact Field Point 2120 as remote target whit some I/O card, and a PC used to releave data and memorize them (host).
    The comunication between the two targets is performed by ethernet lan with Shared Variables not bufferized and not real-time fifo.
    The problem is this: the sistem requires a big number of Vis and Variables (about 20 Vis per target and 50 variables localizated in 5 different librarys all in the remote target), and it should store data, acquisited by the FieldPoint, on a database (SQL Server 2005), one time per minute. This storage is made by Labview Database Connectivity Toolkit, and it work correctly. The problem is an excessive consumption of RAM on the PC.
    The question is this: The use of many shared variables or vis, bring the memory to always grow? I try to explain myself. When we start our application, after the first minute and registration on the database, the memory grow of approximately 1,3 Megabytes. This is repeated every minute, and after approximately two ours, Labview break down and display "Not enought memory".
    We look the on-line help for one solution, and we have found that the variables (local and shared), the concatenate funcion, the array and Vi indicators copy in memory their older value when updated. Is possible that this is the cause of our problem? There is a method to avoid it? How labview uses memory? It is possible set it up for not maintain older value in memory?
    We need to run the application up to 4 days for complete its work, is a long test for a climatic room.
    We're looking forward to hearing from you.
    Best regards
    Michele

    centerbolt is correct, you can't load a .exe or even run a program on the fieldPoint controller unless you have the Real Time module.  However, that does not mean you can't use your FieldPoint bank without the Real Time module. 
    From LabVIEW for windows you can make calls to the fieldpoint IO using the fieldpoint read/write functions. 
    This program runs on the PC not the FieldPoint controller.  If you loose network connection to the fieldpoint, your program will loose connection to the IO.  For many data logging applications this type of arrangement can work just fine.  However, if this is the only type of application you are ever going to run, then you may as well not buy the Real Time controller for your fieldpoint but the network controller only. 
    If your application requires more reliability, and/or greater determinism than can be achieved by running a program on windows, then you should use the LabVIEW Real Time module and develop a program that can run down on the FieldPoint controller independent of windows.
    Message Edited by StevenA on 07-22-2008 04:14 PM
    SteveA
    CLD
    FPGA/RT/PDA/TP/DSC
    Attachments:
    fp pallet.PNG ‏6 KB

Maybe you are looking for

  • Regarding skin and style in obiee 11g

    Hi , i have done all the steps for customization for login page of it. but when i am restarting the presentation server the login page not customized. on the place of oracle logo is coming nothing. i have followed the steps mentione in oracle doc. Pl

  • No audio from logic?

    I have a project that just stopping putting out audio - I have tried unsoloing everything- and nothing.   I was working on a bass part triggering my MOX* Yamaha- which has been working fine-  I set up a track for external midi assigned midi channel 1

  • ITunes Rating Stars Not Solid.

    I've noticed from time-to-time that my rating stars seem to be hollow and no longer solid. What does this happen?

  • Problem Please open the Vi named popup

    Hello I have a problem opening a certain vi when I open a vi lets say "Main.vi" labview popsup a "Please select the vi named: Main.vi" this is a little strange for me. normaly you gets this message if a certain subvi or control cannot be findt in the

  • 64-bit application running SQL Server Compact

    Hi I have read your article sql-server-compact-4-desktop-app-with... I have some questions 1: The app I have is 32/64-bit (installer will detect and install). Will 64-bit app work with SQLCE40? 2: The target framework is .NET 4.0. Will this work? In