Real Time Module, upgrading Labview 6.02 Base Development System?

Hi,
I have Labview 6.02 Base Development System but I need the Real Time module.
Would it be possible to get it upgrading the Labview 6.02? Is the full development systemm needed?
Which could be the cheapest way to get Real Time module?
Thanks.

You can upgrade to the latest version of LabVIEW Real-Time version 7.1 by upgrading both LabVIEW and LabVIEW Real-Time. You can all about the new features of LabVIEW Real-Time at the following link. For more information, please contact us at .
Regards,
Kristi H
National Instruments
>

Similar Messages

  • What are the limitations of using labview 8.5.1 developers suite verses a real-time module in field point applications?

    What are the limitations of using labview 8.5.1 developers suite verses a real-time module in field point applications? Can an exe. be loaded onto a field point controller or does the controlling program have to reside on a PC for example?

    centerbolt is correct, you can't load a .exe or even run a program on the fieldPoint controller unless you have the Real Time module.  However, that does not mean you can't use your FieldPoint bank without the Real Time module. 
    From LabVIEW for windows you can make calls to the fieldpoint IO using the fieldpoint read/write functions. 
    This program runs on the PC not the FieldPoint controller.  If you loose network connection to the fieldpoint, your program will loose connection to the IO.  For many data logging applications this type of arrangement can work just fine.  However, if this is the only type of application you are ever going to run, then you may as well not buy the Real Time controller for your fieldpoint but the network controller only. 
    If your application requires more reliability, and/or greater determinism than can be achieved by running a program on windows, then you should use the LabVIEW Real Time module and develop a program that can run down on the FieldPoint controller independent of windows.
    Message Edited by StevenA on 07-22-2008 04:14 PM
    SteveA
    CLD
    FPGA/RT/PDA/TP/DSC
    Attachments:
    fp pallet.PNG ‏6 KB

  • Is Ardence RTX required when I use PCI-7831R with labview real-time module for RTX target

    I am new in the labview real-time field. I want to use labview real-time module for RTX target on only one desktop. My card is PCI-78331R.  Is Ardence RTX necessary? Why I cannot find my card in the MAX after I configured the RTX?
    Thank you very much.

    Hi Ziaozhongda,
    You must have the RTX runtime engine installed before the LabVIEW Real-Time RTX module will install.  Additionally, you should follow the instructions in the "Using the NI PCI-7831R with the LabVIEW Real-Time Module for RTX" section of Getting Started with the NI 7831R (page 8).
    Please see Configuring Hardware for Use with RTX.
    Also useful is "How Can I Validate That My PC Will Run RTX?

  • Real-Time Module Labview 64 bit

    I have installed labview 2009 64 bit version to my laptop. And, i want to instal the real-time modul to this labview, but i can not do this since it needs labview 32 bit installed. Is it possible for me to install real-time module to 64 bit labview? Please help me with this problem.

    The LabVIEW RT application will most likely always have to remain 32 bit for now. Most of the RT targets are to resource constrained to gain anything from moving them to a 64bit RT OS, in fact in trying to move them to 64 bit they would loose quite a bit of performance. For the Desktop and PXI RT systems, the problem is that they are based on Pharlap ETS and that is a pure 32 bit solution with no plans of ever moving to 64 bit anytime soon, AFAIK. And NI is not likely going to port that to 64 bit themselves, aside from the question if they could technically do so (not sure they have access to all of the source code) and legally (their license most likely does prevent them from changing the ETS software to run under different target systems than the one it was originally purchased for).
    The only thing you can expect to come from NI in the near future is that the RT system also installs into 64 bit LabVIEW, and the actual cross compilation that happens already will simply target the 32 bit RT target. That however means that the RT application itself is still limited by the 32 bit constraints. Maybe there are some plans to add a 64 bit RT system to the mix for the high end PXI controller. What OS that would use is quite unclear. Pharlap ETS seems out of question for that, VxWorks is probably not a favorite solution at the moment and an Linux RT based system has its own legal and technical challenges.
    Rolf Kalbermatter
    CIT Engineering Netherlands
    a division of Test & Measurement Solutions

  • Could I use two regular computer to achieve realtime communicat​ion using the Ethercat and Labview Real Time Module?

    Could I use two regular computers (one acts master , another one acts as slave using the Ethercat Internet Card) to achieve realtime communication using the Ethercat and Labview Real Time Module? if it could, what hardware should I purchase from the NI ?
    Thank you!

    Hi Xiaolin,
    NI doesn't offer Windows based Ethercat master or slave software. Only LabVIEW RT can run the Ethercat driver. 
    However, you could use a LabVIEW RT target as an Ethercat master and use the Ethercat Internet Card with a slave PC (note: I think this will work from the Beckhoff description of the card you are describing. The card should integrate like any other non-NI slave. However I haven't tested the setup and don't advise saying it will work until you have tried). 
    You can use any NI RT target with two ethernet ports as the Ethercat master. This could be a cRIO, PXI or RT Desktop. 
    Jesse Dennis
    Design Engineer
    Erdos Miller

  • The Problem about Monitoring Motion using PCI-7358 on LabVIEW Real Time Module

    Hello everyone,
    I have a problem about monitoring the position of an axis. First let me give some details about the motion controller system I’m using.
    I’m using PCI-7358 as controller and MID-7654 as servo driver, and I’m controlling a Maxon DC Brushed motor. I want to check the dynamic performance of the actuator system in a real time environment so I created a LabVIEW function and implemented it on LabVIEW Real Time module.
    My function loads a target position (Load Target Position.vi) and starts the motion. (Start.vi) then in a timed loop I read the instantaneous position using Read Position.vi. When I graphed the data taken from the Read Position.vi, I saw that same values are taken for 5 sequential loops. I checked the total time required by Read Position.vi to complete its task and it’s 0.1ms. I arranged the loop that acquires the data as to complete its one run in 1ms. But the data shows that 5 sequential loops read the same data?

    Read Position.flx can execute much faster than 5 ms but as it reads a register that is updated every 5 ms on the board, it reads the same value multiple times.
    To get around this problem there are two methods:
    Buffered High-Speed-Capturing (HSC)
    With buffered HSC the board stores a position value in it's onboard buffer each time a trigger occurrs on the axis' trigger input. HSC allows a trigger rate of about 2 kHz. That means, you can store a position value every 500 µs. Please refer to the HSC examples. You may have to look into the buffered breakpoint examples to learn how to use a buffer, as there doesn't seem to be a buffered HSC example available. Please note that you need an external trigger-signal (e. g. from a counter of a DAQ board). Please note that the amount of position data that you can acquire in a single shot is limited to about 16.000 values.
    Buffered position measurement with additional plugin-board
    If you don't have a device that allows you to generate a repetitive trigger signal as required in method 1.), you will have to use an additional board, e. g. a PCI-6601. This board provides four counter/timers. You could either use this board to generate the trigger signal or you could use it to do the position capturing itself. A PCI-6601 (or an M-Series board) provides can run a buffered position acquisition with a rate of several hundred kHz and with virtually no limitation to the amount of data to be stored. You could even route the encoder signals from your 7350 to the PCI-6601 by using an internal RTSI cable (no external wiring required).
    I hope this helps,
    Jochen Klier
    National Instruments

  • End of memory. Labview + Real time module + Compact Field Point 2120

    My name is Michele Ciampichetti, I'm writing from Italy.
    Frist of all, excuse for my english.
    This is my first experience with Labview, and the application is not so easy. We use Labview 8.2.1
    Our system is costituited by two different parts: there is a Compact Field Point 2120 as remote target whit some I/O card, and a PC used to releave data and memorize them (host).
    The comunication between the two targets is performed by ethernet lan with Shared Variables not bufferized and not real-time fifo.
    The problem is this: the sistem requires a big number of Vis and Variables (about 20 Vis per target and 50 variables localizated in 5 different librarys all in the remote target), and it should store data, acquisited by the FieldPoint, on a database (SQL Server 2005), one time per minute. This storage is made by Labview Database Connectivity Toolkit, and it work correctly. The problem is an excessive consumption of RAM on the PC.
    The question is this: The use of many shared variables or vis, bring the memory to always grow? I try to explain myself. When we start our application, after the first minute and registration on the database, the memory grow of approximately 1,3 Megabytes. This is repeated every minute, and after approximately two ours, Labview break down and display "Not enought memory".
    We look the on-line help for one solution, and we have found that the variables (local and shared), the concatenate funcion, the array and Vi indicators copy in memory their older value when updated. Is possible that this is the cause of our problem? There is a method to avoid it? How labview uses memory? It is possible set it up for not maintain older value in memory?
    We need to run the application up to 4 days for complete its work, is a long test for a climatic room.
    We're looking forward to hearing from you.
    Best regards
    Michele

    centerbolt is correct, you can't load a .exe or even run a program on the fieldPoint controller unless you have the Real Time module.  However, that does not mean you can't use your FieldPoint bank without the Real Time module. 
    From LabVIEW for windows you can make calls to the fieldpoint IO using the fieldpoint read/write functions. 
    This program runs on the PC not the FieldPoint controller.  If you loose network connection to the fieldpoint, your program will loose connection to the IO.  For many data logging applications this type of arrangement can work just fine.  However, if this is the only type of application you are ever going to run, then you may as well not buy the Real Time controller for your fieldpoint but the network controller only. 
    If your application requires more reliability, and/or greater determinism than can be achieved by running a program on windows, then you should use the LabVIEW Real Time module and develop a program that can run down on the FieldPoint controller independent of windows.
    Message Edited by StevenA on 07-22-2008 04:14 PM
    SteveA
    CLD
    FPGA/RT/PDA/TP/DSC
    Attachments:
    fp pallet.PNG ‏6 KB

  • Real Time Module needed for hi speed digitizer PCI 5911 ?

    Rookie question :
    I run LV 7.1 , LV Run Time  8 ( whatever this Run Time Engine may be ) , MAX  4.0.0 , NiDaq 8 ( for my two Multi Function PCI Daq's ).
    Question : does the PCI 5911 work under NI-Scope 3.3 ( supporting PCI 5911 ) without a Real Time Module ?
    i.e. is the Real Time Module a separate ( hardware ? ) device, not necessarily needed for PCI 5911 ?  
    Sorry for this utter basic question.
    I have not used a hi speed digitizer before, only multifunction NiDaq's.
    Thanks for any help or reference.

    You are probably referring to the LabVIEW Real-Time Module described here.
    This is an extension to LabVIEW that allows you to run VIs on a
    real-time, deterministic environment, usually on remote hardware.  One
    common configuration is to use a Windows system as your host, and
    communicate to a remote PXI controller running LabVIEW RT. 
    The PCI-5911 is fully supported for "standard" LabVIEW running on Windows.  You don't need LabVIEW RT to use it.
    In fact, the PCI-5911 is not supported under LabVIEW RT.
    Message Edited by psisterhen on 10-16-2007 09:50 AM

  • Is real time module necessary?

    Hi,
    I want to use cRIO-9073 for simple system control. Do I need Labview Real-time module if I don't need real-time support?
    Thanks in advance.
    Karpis

    Hi Karpis,
    This is the forums for IF-RIO, which is the PCI-5640R and PXIe-5641R devices. Please post your question to the cRIO forums found here:
    http://forums.ni.com/t5/Real-Time-Measurement-and/bd-p/280
    You can get faster response to your question there.
    Best Regards,
    Jignesh P
    Applications Engineer

  • Real-Time Module architecture

    Hi, (I use LV RT DSC 8.2.1)
    I want to learn how to use the Real-Time Module.
      I have process that is running at a Real Time compact Field Point Target.
    The process should be controlled and monitored by the Target but the triggering of the process should be available at the Host.
    I also want to view the Process progression at the Host screen.
      As far as I understand it I have several options :
    Create a RT FIFO buffer and send the process data to that buffer, and then, use 'TCP Write' to send the buffer data to the Ethernet Network, where it is available to any Host PC at the Ethernet Network to view (by using TCP Read)
    Create a Variable in the Project, and enable his RT FIFO at his properties.
    Control the process at the Target with Shared Variables and read them at the Host.
      Option 3 is the option I have been using so far, ignoring the RT property of the controllers, what are the advantages of using RT module ?
    and when should I consider using them ?
      What is the difference between 1 and 2 and what are their advantages ?
      What important things do I need to consider when I use RT Module  (just several tips if possible) ?
       Sincerely Yours,  Amitai Abramson.

    Hi Amitai,
    You could go through this link to know more about RT FIFO and its use. http://zone.ni.com/devzone/cda/tut/p/id/3934
    and this for communication wizard http://zone.ni.com/devzone/cda/tut/p/id/4676
    Message Edited by mallyaprs on 04-30-2009 03:53 AM
    -Mallya
    " Appreciate and motivate those who help"

  • NI LabVIEW Base Development System for Windows and NI CVS-1454

    I have NI CVS-1454 and NI LabVIEW Base Development System for Windows (LabVIEW 8.5, no LabVIEW Real-Time Module and Vision Development Module). How can I build a LabView-program and run it on my CVS-1454?
    Kirill Kostousov

    Hi Kirill,
    In order to use your CVS-1454 you will need to buy the LabVIEW Real-Time Module and Vision Development Module as these features are not supported in your version of LabVIEW.
    Thomas Clark
    Applications Engineer
    National Instruments UK & Ireland

  • Low pass filter in LabVIEW Base Development System

    Hello,
    I've got a problem for several days and I haven't find even a hint to fix it so it's driving me crazy.
    I'm developing an application on LabVIEW Base Development System for automatical aquisition and exploitation of five signals.
    These signals are generated by a deformation gauge and four potentiometers linked to a NI USB 6210.
    I'm using the DAQ assistant on LabVIEW to get and use these signals.
    My problem is that I need a 2nd order low pass filter on the deformation gauge signal to get something I can use afterward and none is include in the Base development Sytem of LabVIEW.
    I'm sure there is a way to pass through this without buying the Full Development version but I haven't found it yet.
    Does anybody has already had this problem and succeed in fixing it ?
    If yes, can I have the code ?
    Thanks

    Thanks for your responses.
    I've tried to create a subVI using the transfer function of the filter I need.
    Here is the diagram of this subVI :
    I'm not sure of this transfer function at all. It seems too simple to me.
    Does it look weird to you or does a low-pass second order filter can have such a simple transfer function ?
    I've tried it and my signal is still unstable. I guess I'm doing something wrong, but as I'm not that good in electronics, I can't figure out what.
    Thanks again

  • LabVIEW Base Development System for Windows

    Hello,
    I want to buy "LabVIEW Base Development System for Windows", but i have some questions, include this version functi...
    a)TDMS streaming VIs
    b)Wavefoem VIs
    and the last question, is it possible with this version do some executable code?
    thanks for all..

    Here is a decent list of the differences between the versions of LabVIEW.  Since you want to create executables, you will need to add the Application Builder.  If you find there are other functions that you may need, I would just bump up to the Professional version.
    There are only two ways to tell somebody thanks: Kudos and Marked Solutions
    Unofficial Forum Rules and Guidelines

  • Memory Footprint for LabVIEW 7.1 Full Development System

    Hello,
             I would like to know the memory footprint for LabVIEW 7.1 Full Development System.Both during the installtion and for developing applications later.
    Can someone help.Thank you.
    Solved!
    Go to Solution.

    There isn't a single answer for that, as this varies considerably.
    As far as disk space goes, if memory serves, the basic 7.x takes up a few hundred MB, but that could inflate considerably depending on which modules and drivers you add to it.
    For RAM, I believe 7.0 used to take up ~20-50 MB immediately after being loaded, but that number would then also change a lot over time depending on what you do. Loading, editing and running code all regularly change the amount of RAM used.
    Why are you asking? Also, are you aware that 7.1 is relatively old by now (came out 7 years ago)?
    Try to take over the world!

  • How can I display a .csv file as a real time waveform using labview

    Hi folks.
    As stated, Im looking to display a .CSV file (obtained via RS-232, example; 100,101,102,102,102,103...) as a waveform in real-time. What is the best way to approach this? Im thinking LabVIEW. I have version 7.1 but am not really experienced in using it. I'll give anything a go though. Also, I have Visual Studio 2005, so I was thinking VB???
    Thanks in advance!
    Simon.
    Message Edited by simonious on 10-26-2007 04:13 AM

    Have a look at this VI.
    Add a convert string to number function behind it.
    André
    Message Edited by andre.buurman@carya on 10-26-2007 01:08 PM
    Regards,
    André
    Using whatever version of LV the customer requires. (LV5.1-LV2012) (www.carya.nl)
    Attachments:
    convert csvars 2 array of vars.vi ‏16 KB

Maybe you are looking for

  • Activation of "Convert to sRGB" in "Save for Web"

    Hello, I am using a script scaling and exporting my image via "Save for Web and Devices". When I open the menu there's a checkbox called "Convert to sRGB" - I want to activate that checkbox via Script, as then I get the wanted colour results. I alrea

  • RGB vs. CMYK Colorspace - Which is best to work in?

    I'm designing a document for a client, within in InDesign. Up until my last save, I've been ignoring the notice about InDesign being set in the RGB colorspace. On my last save, I changed it to the CMYK colorspace and all my linked images got lighter

  • 5530 (probably) temperature problem

    I have 5530  an year and a few months already. Approximately at second month it began rebooting automatically . Firstly thought that it's some software problem. ( Now I'm not sure is it really)So I found that it reboots only when device temperature d

  • Calling backing bean method from HttpSessionListener

    Hi, I am detecting session expiration through HttpSessionListener's sessionDestroyed() method. I have a logout() method in my backing bean which does all the session clean up. If the user clicks the logout link, then I call the logout method and do t

  • KDE crash (and pannels disappear) after update

    Just now I updated, and KDE was updated to 4.5.3-1. After that update, kwin crashed, and both desktop widgets and panels disappeared. I restarted X and that brought back the widgets, but not the panels... Which is very much annoying, because besides