Keithley 2400 Real Time Resistance Values in LabView 2010

Hi,
I keep getting Error 803 and Error -230 on my Keithley 2400 SourceMeter (connected NI-GPIB-USB-HS; everything is talking correctly) when I use the attached LabView program. I'm using the most updated drivers for VISA, 488.2, and the KE24XX. I've looked at the VI Tree and everything should be okay. Individually, the VIs are not causing errors. Also important to note is that when I take 4W resistance readings from the Keithley without the PC interface the Keithley's display indicate different settings, including ARM, while when recording from computer it displays AUTO and 4W. No data is displayed on the device while recording from computer, while data is displayed without recording from the computer. 
Please let me know what you think. Thank you!
Attachments:
resistdatawrite.png ‏217 KB

Hi coldcarbon!
I did a search and it looks like this is not a totally uncommon occurrence with these devices.  I found a few forums searching google, and also was able to find a discussion forum at Keithley's website.  These drivers look like they were built by Keithley, and I would encourage you to get in touch with them to figure out what exactly is going on.  I wouldn't jump to that issue being a hardware problem just yet.
Get back in touch with us after you talk to Keithley and we might be able to help you more with this issue, but it is probably an issue with the software, and Keithley would definitely be the people to talk to.
Ben N.
Applications Engineering
Certified LabVIEW Developer

Similar Messages

  • How to measure and plot RMS value of real time data?

    hi
    i need to measure and plot real time RMS value of EMG voltage. I made a VI. But I dont know why it didn't work. Can anyone please help me out? My VI is attached  in 2013 and 2011 both version. One sample data is also attached. Thanks at advance.
    Solved!
    Go to Solution.
    Attachments:
    RMS.vi ‏47 KB
    RMS_2011.vi ‏41 KB
    Sample Data.docx ‏20 KB

    You seem to select EMG voltage vs Angle(Z-axis). Is that really what you want? More resasonable would seem e.g. EMG voltage vs. time.
    If you want to process the 3D (xyz data), you probably need to add some more math). Why are you taking the absolute value? RMS does not care about the sign anyway, right?
    They you are trying an XY graph with a single point, because your Y value is only a single number. You need an equal number of points for Y.
    Here's a quick modification that graph EMG(mV) vs. time(ms) and takes the RMS of the EMG data. See if this gives you some ideas.
    Please provide more information what you want to do with the angles.
    LabVIEW Champion . Do more with less code and in less time .
    Attachments:
    RMS_MOD.vi ‏18 KB

  • End of memory. Labview + Real time module + Compact Field Point 2120

    My name is Michele Ciampichetti, I'm writing from Italy.
    Frist of all, excuse for my english.
    This is my first experience with Labview, and the application is not so easy. We use Labview 8.2.1
    Our system is costituited by two different parts: there is a Compact Field Point 2120 as remote target whit some I/O card, and a PC used to releave data and memorize them (host).
    The comunication between the two targets is performed by ethernet lan with Shared Variables not bufferized and not real-time fifo.
    The problem is this: the sistem requires a big number of Vis and Variables (about 20 Vis per target and 50 variables localizated in 5 different librarys all in the remote target), and it should store data, acquisited by the FieldPoint, on a database (SQL Server 2005), one time per minute. This storage is made by Labview Database Connectivity Toolkit, and it work correctly. The problem is an excessive consumption of RAM on the PC.
    The question is this: The use of many shared variables or vis, bring the memory to always grow? I try to explain myself. When we start our application, after the first minute and registration on the database, the memory grow of approximately 1,3 Megabytes. This is repeated every minute, and after approximately two ours, Labview break down and display "Not enought memory".
    We look the on-line help for one solution, and we have found that the variables (local and shared), the concatenate funcion, the array and Vi indicators copy in memory their older value when updated. Is possible that this is the cause of our problem? There is a method to avoid it? How labview uses memory? It is possible set it up for not maintain older value in memory?
    We need to run the application up to 4 days for complete its work, is a long test for a climatic room.
    We're looking forward to hearing from you.
    Best regards
    Michele

    centerbolt is correct, you can't load a .exe or even run a program on the fieldPoint controller unless you have the Real Time module.  However, that does not mean you can't use your FieldPoint bank without the Real Time module. 
    From LabVIEW for windows you can make calls to the fieldpoint IO using the fieldpoint read/write functions. 
    This program runs on the PC not the FieldPoint controller.  If you loose network connection to the fieldpoint, your program will loose connection to the IO.  For many data logging applications this type of arrangement can work just fine.  However, if this is the only type of application you are ever going to run, then you may as well not buy the Real Time controller for your fieldpoint but the network controller only. 
    If your application requires more reliability, and/or greater determinism than can be achieved by running a program on windows, then you should use the LabVIEW Real Time module and develop a program that can run down on the FieldPoint controller independent of windows.
    Message Edited by StevenA on 07-22-2008 04:14 PM
    SteveA
    CLD
    FPGA/RT/PDA/TP/DSC
    Attachments:
    fp pallet.PNG ‏6 KB

  • To develop an VI to aquire data of thermocouples from a remote chamber in labVIEW real time, should use NIDAQ or NIDAQmx?

    Hai
    I need to develop an VI to aquire data at real time. Am using LabVIEW RT as the OS and i need to know, if i can use NIDAQ 7.0 or should i use NIDAQmx? what is the diffrence between the two? is it just that NIDAQmx makes job simpler?
    Also can anybody please tell me, how to access a channel in labVIEW? am not clear about the path to follow to get the physical channel as a virtual channel on the LabVIEW.
    I am using a 8186 PXI with labVIEW RT and a SCXI, with both scxi and daq card on the same chassis.
    the development will be done on Windows XP which will be the host machine.
    Thanx in advance for any answers
    Arun

    Arun,
    DAQmx is our new Data Acquisition driver and should be considered as the primary option when writing new applications because of supportability in the future.
    DAQmx does make Data Acquisition tasks much simpler, abstracting many of the board implementation details and giving you a common API for 99% of the tasks you might want to accomplish.
    There are a couple of minimum requirements for using DAQmx on LabVIEW Real Time:
    - You will need LabVIEW Real-Time 7.1 (or higher)
    - You will need DAQmx 7.2 (or higher)
    These requirements are mentioned in the following Knowledgebase:
    Is NI-DAQmx Supported in LabVIEW Real-Time?
    The other thing you might want to make sure of is that DAQmx supports the DAQ hardware you already have or are planning to use. DAQmx does support most of the existing DAQ and SCXI hardware products, and reading thermocouples should become a very simple task.
    With DAQmx you can use physical channels, Global Channels (Aliases for preconfigured physical channels, equivalent to the Traditional DAQ Virtual Channels) or DAQmx Tasks (Preconfigured Data Acquisition tasks that include channel, timing and triggering information). A good place to start to learn about these features could be:
    Advantages of NI-DAQmx
    The following link might be useful to check Hardware support:
    http://www.ni.com/support/daq/versions.htm
    The driver can be downloaded by following these links:
    ni.com -> Support -> Drivers and Updates -> Current Software Versions -> Multifunction DAQ -> NI-DAQ Version 7.3 for Windows 2000/NT/XP
    This includes the latest version of DAQmx and our Traditional DAQ driver. Please note that DAQmx base is also listed but that's our Register-Level Programming (RLP) driver, which provides DAQ support for previously unsupported platforms.
    I hope this helps,
    Alejandro

  • What are the limitations of using labview 8.5.1 developers suite verses a real-time module in field point applications?

    What are the limitations of using labview 8.5.1 developers suite verses a real-time module in field point applications? Can an exe. be loaded onto a field point controller or does the controlling program have to reside on a PC for example?

    centerbolt is correct, you can't load a .exe or even run a program on the fieldPoint controller unless you have the Real Time module.  However, that does not mean you can't use your FieldPoint bank without the Real Time module. 
    From LabVIEW for windows you can make calls to the fieldpoint IO using the fieldpoint read/write functions. 
    This program runs on the PC not the FieldPoint controller.  If you loose network connection to the fieldpoint, your program will loose connection to the IO.  For many data logging applications this type of arrangement can work just fine.  However, if this is the only type of application you are ever going to run, then you may as well not buy the Real Time controller for your fieldpoint but the network controller only. 
    If your application requires more reliability, and/or greater determinism than can be achieved by running a program on windows, then you should use the LabVIEW Real Time module and develop a program that can run down on the FieldPoint controller independent of windows.
    Message Edited by StevenA on 07-22-2008 04:14 PM
    SteveA
    CLD
    FPGA/RT/PDA/TP/DSC
    Attachments:
    fp pallet.PNG ‏6 KB

  • Real-Time Module Labview 64 bit

    I have installed labview 2009 64 bit version to my laptop. And, i want to instal the real-time modul to this labview, but i can not do this since it needs labview 32 bit installed. Is it possible for me to install real-time module to 64 bit labview? Please help me with this problem.

    The LabVIEW RT application will most likely always have to remain 32 bit for now. Most of the RT targets are to resource constrained to gain anything from moving them to a 64bit RT OS, in fact in trying to move them to 64 bit they would loose quite a bit of performance. For the Desktop and PXI RT systems, the problem is that they are based on Pharlap ETS and that is a pure 32 bit solution with no plans of ever moving to 64 bit anytime soon, AFAIK. And NI is not likely going to port that to 64 bit themselves, aside from the question if they could technically do so (not sure they have access to all of the source code) and legally (their license most likely does prevent them from changing the ETS software to run under different target systems than the one it was originally purchased for).
    The only thing you can expect to come from NI in the near future is that the RT system also installs into 64 bit LabVIEW, and the actual cross compilation that happens already will simply target the 32 bit RT target. That however means that the RT application itself is still limited by the 32 bit constraints. Maybe there are some plans to add a 64 bit RT system to the mix for the high end PXI controller. What OS that would use is quite unclear. Pharlap ETS seems out of question for that, VxWorks is probably not a favorite solution at the moment and an Linux RT based system has its own legal and technical challenges.
    Rolf Kalbermatter
    CIT Engineering Netherlands
    a division of Test & Measurement Solutions

  • Qual a diferença do LabVIEW FPGA e o LabVIEW REAL- TIME?

    Hello,
    Could anyone help me? This question arose in my work and could not answer. Does anyone know tell me?
    Thank you.
    Solved!
    Go to Solution.

    Olá, 
    Com o LabVIEW FPGA você programa diretamente o chip FPGA presente no chassis cRIO, placas da NI Série R e single Board RIO. Você programa as E/S diretamente no chip, sendo assim uma programação de mais baixo nível. A grande vantagem do FPGA é o "paralelismo real", o que garante altas velocidades na execução das rotinas programadas.
    Assista ao webcast Introduction to LabVIEW FPGA
    Com o LabVIEW Real-Time você desenvolve aplicações que são críticas em relação a "tempo". Chamamos de aplicações "deterministicas". A idéia é que, se você tem uma rotina que deve ser executada em um tempo determinado, a diferença entre o tempo real e o que você programou seja o menor possível. Aplicações em Computadores comuns não são deterministicas, pois enquanto o LabVIEW executa uma certa rotina programada, o Windows está executando outras tarefas, monitorando a utilização de periféricos, atualizando a tela, etc.
    Real time não significa "Velocidade", mas "Confiabilidade".
    Assista ao webcast LabVIEW Real-Time: Graphical Development, Hard Real-Time Performance
    Hello, 
    With LabVIEW FPGA you program the FPGA chip itself. The FPGA chip is found into cRIO chassis, NI R-Series Boards, and NI Single board RIO. You program the I/O directly in the chip, so we consider this as a low level programing. The main FPGA advantage is the "Real Paralelism", which guarantees high speed execution programming.
    Whatch Introduction to LabVIEW FPGA webcast
    With LabVIEW Real-Time you develop called "time crictical" or "Deterministics" applications. In case of you routine must be executed in a specified period of time, the difference between the real time execution and the time you programmed is as low as possible. Commom computers programming are not deterministic, because in the meantime of executing a certain programmed routine, the OS (e.g. Windows) is running other appication tasks, monitoring peripherals like mouse and keyboard, uptading screen, etc.
    Real-Time doesn't mean "faster" but "reliable".
    Watch LabVIEW Real-Time: Graphical Development, Hard Real-Time Performance webcast
    I hope the information helps you!
    Best Regards
    Felipe Flores
    Engenharia de Aplicações
    National Instruments Brasil

  • HOW CAN I ACQUIRE AND DEPLOY DATA IN REAL TIME WITH DAQ 6024E?

    HI ENGINEER,
    I'M PROGRAMING AN APLICATION WHICH I NEED ACUIRING AND DEPLOYING AN ANALOGOUS SIGNAL IN REAL TIME.
    I HAVE:
    LabVIEW 7.0
    WINDOWS 2000
    DAQ CARD 6024E
    THE SIGNAL I WANT TO ACQUIRE IS 0-10 Vcd FROM A SENSOR, THE PROBLEM I HAVE, IS THAT I AM ACQUIERING THE DATA AND IT SEEM THAT THEY ARE UPDATING EACH SECOND, BUT I HAVE BEEN MOVING THE PARAMETERS SUCH AS TASK TIMING IN THE DAQ ASSISTANT VI AND NOTHING HAPPENS.
    CAN YOU TELL ME WHICH PARAMETERS I HAVE TO MOVE TO AQUIRE AND DEPLOY DATA IN REAL TIME

    Hi barron,
    If you want to see the data update quicker, you can decrease the value in the "Samples to Read" field of your DAQ Assistant. This number controls when data is transferred from the DAQ buffer to the PC. If the sampling frequency remains the same, but this field decreases, it means the buffered data will be returned more quickly.
    Hope this helps.
    Thanks,
    Lesley Y.

  • Real-time pid control

    Hi,
    I found a real-time PID control in LabView examples, for E Series boards. I used one of this boards a time ago and used this control. It works very well.
    Now, I have a M series board (PCI-6221, LabView 7.0) and I would like to use a similar program. I tried to use the DAQmx assistant and look at the examples, but I did not find anything similar.
    Could you please take a look in this example (attached) and help me in configuring the PCI-6221 to work in similar way?
    I really apreciate your support.
    Thanks,
    Jeferson
    Attachments:
    Real-Time PID Control2.vi ‏131 KB
    simple PID.vi ‏53 KB

    Jeferson,
    I no longer have LabVIEW 7.0 on my machine, but I know that in LabVIEW 7.1, DAQmx installs an example called "PID Control-Single Channel.vi." This should be exactly what you are looking for. It can be found in the LabVIEW Example Finder (Help->Find Examples) by searching for "PID."
    Basically, this VI uses hardware timing (the rate will depend on your computer specifications) to acquire a single point, calculate the PID output, then output that value on the next sample clock.  Please let us know if you have any additional questions!
    Hope this helps,
    Ryan Verret
    Product Marketing Engineer
    Signal Generators
    National Instruments

  • How to get real time date from Sunny Boy Inverters through SMA Webbox?

    Hello,
    I was looking on how to communicate with three Sunny Boy Inverters that are linked to a SMA webbox. Currently we get a daily summary of the solar production, but we would like to have real time data uploaded into LabVIEW. The people of SMA said that I needed to use "RS485 protocols" with a software they have called YASDI (http://www.sma.de/en/products/software/yasdi.html); however, I keep finding other information that says that I can communicate with my inverters using the SMA OPC server (http://www.sma-america.com/en_US/products/software/sma-opc-server.html). I also found this forum, http://forums.ni.com/t5/LabVIEW/Sunny-Island/m-p/809129/highlight/true#M368718, where I understood that I don't need to use the OPC server, but instead I can use UDP communication to get the data. I found another forum that talks about using the YASDI :http://forums.ni.com/t5/LabVIEW/help-with-Sunny-Boy-inverter/m-p/293273/highlight/true#M153892. 
    I've tried to communicate with the people of the forums and I haven't gotten any responses; therefore, I wanted to see if I can get cleared up on this issue. Do I need to use OPC server to communicate with my inverters through the webbox or can I just use UDP in LabVIEW?
    Any help would be greatly appreciated,
    Thanks
    Jose

    Hello,
    I got a response from Vincezo: http://forums.ni.com/t5/LabVIEW/Sunny-Island/m-p/1794786. I'm currently working with the VI to communicate with the webbox through UDP. Since the PRC protocols to communicate with the webbox are in JavaScript Object Notation (JSON), I'm not sure if the data-in string has to be in JSON format. However, the only error I'm getting is 56 in the UDP read function. I've tried increasing the size of the bytes and timeout, but I keep getting this error. If anyone can help me with this part I would very grateful. 
    Thanks,
    Jose

  • LabVIEW2013,LabVIEW2013SP1(Real-Time付)インストールができない

    LabVIEW2013,LabVIEW2013SP1(いずれもReal-Time付)インストールの件で困っています。
    NIの技術サポートにも数度問い合わせていますが、解決していないので、 皆様のお知恵を貸して頂ければと思い、投稿します。
    経緯
    2月末にLabVIEW2013(Real-Time付)を購入し、下記パソコンにインストールを行いました。
    しかし、インストール中、度々エラーメッセージが表示されておりました。
    パソコン: HP Compaq Elite 8300 US/CT
    OS: Windows7 32bit PRO
    パソコンも最近買ったものです
    その後、LabVIEW2013SP1(Real-Time付)のDVDが届いたので、これをインストールしたところ、
    やはりエラーメッセージが表示され、インストールが完全に成功していないようでした。
    ただ、確認のため、同時に購入したPXIをMAXで認識させて、 Real-Timeのプロジェクトを作る準備ができるかなと思い、付属の手順のとおり作業しましたが、 Real-Timeのプロジェクトが作れませんでした。
    (手順書のとおりできない・・・RealTimeに関する項目が出てこない)
    また、サンプルファインダでReal-Timeのプログラムを探して開きましたが、絵のとおり!マークが表示されます。
    インストールがうまくいっていないと思い、 アンインストールとインストールを数度繰り返しています。
    最初のLabVIEW本体のインストールは成功しますが、そのあとのReal-Time以降インストールが失敗し、 あとはエラーメッセージが表示され、ほとんどインストールできません。 何度か再インストールを行っていますが、いまだ解決に至っていません。
    再インストールの際、行ったことは、
    1.HP Protect Tools一式の削除(hp制のセキュリティソフト)
    2.Intel Management and Security Statusの終了
    3.ファイアウォールの無効化
    です。
    なお、3のファイアウォールの無効化の際、ドメインプロファイルのファイアウォールも無効にしたところ、 再インストールの際の動作が改善し、その前よりもインストールできたソフトが多くなりました。
    また、インストール作業の最後のほうで、「ファイアウォールが有効になっています」と毎回表示されます。
    (ファイアウォールは無効にしているにも関わらず)
    質問:
    1.インストールに失敗する原因と解決策のヒントがありましたら、教えてください。
    2.上記パソコンでのインストール成功例がありましたらお知らせください。
    参考までに絵を添付します。 エラーメッセージはいくつも出ていますが、抜粋して添付します。
    インストールの概要
    インストールされていない
    NI SignalExpress 2013
    NI Real-Time Execution Traceツールキット2013
    LabVIEW 2013 VI Analyzerツールキット
    LabVIEW 2013 Report Generation for Microsoft Officeツールキット
    LabVIEW 2013 PID and Fuzzy Logicツールキット
    LabVIEW 2013 Unit Test Frameworkツールキット
    LabWindows/CVI 2013 SP1 開発システム
    NI デバイスドライバ - 2014年2月版
    インストール済み
    LabVIEW 2013 SP1 (日本語版) (アクティブ化)
    LabVIEW 2013 SP1 Real‐Timeモジュール - 日本語版       ←インストール済になっていますが、実際は使用できません LabVIEW 2013 Digital Filter Designツールキット (アクティブ化)
    LabVIEW 2013 Database Connectivityツールキット (アクティブ化)
    LabVIEW 2011 Advanced Signal Processingツールキット (アクティブ化)
    LabVIEW 2013 DataFinderツールキット (アクティブ化)
    LabVIEW 2013 Desktop Execution Traceツールキット (アクティブ化)
    LabVIEW 2013 System Identificationツールキット (アクティブ化)
    LabWindows/CVI SQLツールキット2.2 (アクティブ化)
    LabWindows/CVI Signal Processingツールキット7.0.2 (アクティブ化)
    LabWindows/CVI PID Controlツールキット2.1 (アクティブ化)
    LabWindows/CVI Execution Profiler 1.0 (アクティブ化)
    NI Real-Time Execution Traceツールキット2013 (アクティブ化)
    NI一般セキュリティパッチQ2 2013
    解決済!
    解決策の投稿を見る。
    添付:
    サンプル.PNG ‏73 KB
    13411.PNG ‏115 KB
    ドメインプロファイル2.png ‏229 KB

    まずReal Timeについて確認させてください。
    インストール済み
    LabVIEW 2013 SP1 (日本語版) (アクティブ化)
    LabVIEW 2013 SP1 Real‐Timeモジュール - 日本語版
    LabVIEW 2013 Digital Filter Designツールキット (アクティブ化)
    の中で、Real-Timeモジュールだけがアクティブ化となっていないことが気になります。
    一応の確認ですが、License ManagerでReal Time Moduleはどのような状態になっていますか?
    アクティブ化されていれば添付のように緑色のアイコンになっているはずなのですが
    Tepig
    Certified LabVIEW Associate Developer
    添付:
    rt.PNG ‏2 KB

  • Cannot communicat​e to Keithley 2400

    I wrote a small Vi for using Keithley 2400 to measure Voltage based on labview 6.2. After 10 month, I found that my program cannot talk to the Keithley 2400. I do not know becasue I upgrade Labview to labview 8.2, or someone deleted some key drivers for GPIB in this public computer. I attached the Vi.   Thanks so much for help.     
    Attachments:
    volttest2.vi ‏81 KB

    See the red border around the whole edge of your block diagram? That's the breakpoint that's sorrounding your entire block diagram. Breakpoints are shown by either big red circles if they're on a wire, or a red border around a VI/function if it's set on a VI/function. In this case you must have accidentally clicked on the block diagram.
    Have you taken any of the LabVIEW tutorials? To learn more about LabVIEW it is recommended that you go through the introduction material, tutorial(s), and other material in the NI Developer Zone's Learning Center which provides links to other materials and other tutorials. You can also take the online courses for free.

  • How to set the time difference between each data when sweeping voltage using keithley 2400

    Hello friends,
    I am using the sweep vi from Keithley in the SWEEP and Acquire measurement vi.I want to measure the voltage for every step and insert a delay in between every two data, so I need some delay between each I step.
    I am a starter to use Labview, thanks so much for your answers.
    Perry
    Solved!
    Go to Solution.
    Attachments:
    Keithley 24XX Sweep and Acquire Measurements.vi ‏26 KB

    As Dennis said, if you're using the built-in sweep function, you'll have to check the manual. See Section 10-16 (That's section 10 page 16, not sections 10 though 16) of the Keithley 2400 User Manual.
    The Keithley 24xx series has a measurement speed in units called PLC (Power Line Cycles). The default speed is 1PLC, meaning a measurement is taken every 1 power line cycle, or 1/60th of a second (16.67ms). 24XX's can go from 0.01 PLC (every 0.16ms) to 10 PLC (every 166.6ms). The faster you measure, the lower precision you get.
    To set this value programmatically, the command is
    ENSe:CURRent:NPLCycles <n>
    ENSe:VOLTage:NPLCycles <n>
    depending on what you're sensing and where <n> is the number of PLCs from 0.01 to 10.
    Another factor that will determine the time between data points is the SDM cycle. These are more complicated, look at your Keithley user manual for more information. Look at Section 6 and Section 11 for more info.
    Note:
    PLC times are based on a US cycle of 60Hz.

  • More than 1 ms MOMO digital IO 7358 in LabView Real-Time 8.5, error downloading EPOS dll file into RT

    Hi all,
    I have several problems here, I tried to google but still I could not solve the problem.
    Here are my questions:
    1. I tried to download an example program of EPOS which require to download a DLL file into the RT, but it failed, here's the error (I attached the picture) 01/02/2008 13:29:50.511 [error] LabVIEW:
    Failed to load shared library EposCmd.dll:_VCS_GetProtocolStackName@16:C on RT target device.
    From the NI.com forum,
    I found a similar problem and it said by installing Network Variable Engine and Variable Client Support into the LV-RT, and I did, but still failed. Any one can help me about this?
    2. I posted a question also in lavag, I think this section is more appropriate. "I
    am trying to use the digital IO of the motion controller of 7358, but then I realized the delay is more that 1 ms (I checked using the oscilloscope), I attached the code (MOMO = Must On Must Off). Is this real, I mean the hardware delay between on-off is as big as this? Anyone knows the solution for this?"
    Note: I am using LV-RT 8.5, PXI-7358 Motion Controller, EPOS 70/10
    Thank you for any help
    Message Edited by bono02 on 01-02-2008 05:47 AM
    Attachments:
    maxon-rt3.JPG ‏120 KB
    momo1.JPG ‏55 KB

    Hello bono,
    1.  From what I have seen, this error is usually caused by one of two things.  The first is the Shared Variable software components not being installed, which is covered in this Knowledgebase.  The other is due to a corruption of the ni-rt.ini file on the Real-Time controller.  You can follow the steps layed out in this Knowledgebase, which gives three possible remedies.  You could also try reinstalling the software (firmware) or create a PXI Uninstall disk from Measurement & Automation Explorer by selecting Tools»Remote Systems»RT PXI Disk Utilities»Create PXI Uninstall Disk.  Boot the PXI Controller from the PXI Uninstall Disk, which will erase the ni-rt.ini and ph_exec.exe on the target.  However, you will need to reconfigure the controller after doing this.
    2.  Due to the nature of the digital I/O on motion controllers, it is not surprising that you are seeing around a 1ms delay.  It deals with the fact that it must update the motion controller with the new value on each control loop.  This Knowledgebase gives a little more clarification regarding the 7350 specs.
    Carlton
    CLD

  • How to create a c++ shared library (.so) for linux real time (for myRio) with Eclipse to use in LabView?

    I tried already these Tutorials and Advices but I didn't find a solution:
    - http://www.ni.com/tutorial/14625/en/
    - http://www.ni.com/tutorial/14690/en/
    - http://forums.ni.com/t5/LabVIEW/Shared-Library-on-myrio-Linux-Real-time-system/m-p/2842540/
    - http://forums.ni.com/t5/LabVIEW/How-to-create-shared-library-for-linux-real-time-target-in/m-p/28218...
    - and some more
    I want use c++ codes on linux real time. For testing reasons I want to have a function that adds 2 values and gives the result.
    I've done these steps:
    1. writing a c++ file in Eclipse (see screensot 2)
    2. building a shared library (.so) from my c++ project in Eclipse (with Cross GCC)
    3. putting this file on myRio (path: /usr/local/lib/)
    4. creating a VI that calls this library from Labview with a "Call Library Function Node" (see screenshot3)
    5. Setting the properties for the "Call Library Function Node" (see screenshots 4-7)
    After I run this VI i get this error message: LabVIEW:  (Hex 0x627) The function name for the ... node cannot be found in the library. To correct this error, right-click the Call Library Function Node and select Configure from the shortcut menu. Then choose the correct function name. (see screenshot1)
    I've tried a lot things to solve this problem but I couldn't find a solution. Would be very happy if anyone can help me. I guess that I have to edit my c++ code to export my function (symbol). But I have no idea how to make it. I also tried it with a dll file in the same folder but it didn't help.
    Perhaps someone can send an example which works on myRIO.
    Thanks!
    screenshot1
    screenshot2
    screenshot3
    screenshot4
    screenshot5
    screenshot6
    screenshot7

     can see it in the screenshot8 there is a function called "_Z8AddierenddPd" instead of "Addieren". I copied this name to Labview (see screenshot9) and it worked.
    I'm sure that there is a way to compile the shared folder with gcc without decorations (mangling). But I don't know how. If someone has a recommendation I would be very glad!
    Prepend each function declaration that you want to be available without name decoration with
    extern "C" <your function declaration>
    Or if you have multiple functions you want to export you can in the header file where you declare your functions simply use:
    #ifdef __cplusplus
    extern "C" {
    #endif
    <all your function declarations>
    #ifdef __cplusplus
    #endif
    Rolf Kalbermatter
    CIT Engineering Netherlands
    a division of Test & Measurement Solutions

Maybe you are looking for