Difficulty synchroniz​ing data from VISA resource with data from a physical channel

Essentially what I'd like my program to do is control the electrical power going (sourcing either current or voltage) into a light and measure the intensity of the light at a given power level, and then do this automatically for ~1000 increments of the source voltage/current.
One of my lab partners made a program a while back which does what we want("LIV Sweep Rev.vi" the first image in the link at the end of this post), but it's a bit sloppy: the program is able to interact with the power supply (connected directly to the computer via USB) and make it turn on, increase the voltage/current, and record the "IV" characteristics just fine. The program can also interact with the light detector (connected via NI-BNC 2110) and gather the photocurrent data. The problem, however, is that the data wouldn't be in sync. The photocurrent data for when the light was actually supplied with 1V would be improperly recorded in the cell for when 2V are applied to the light. To fix this problem my lab partner added in a time delay so that the detector will pause for ~0.35 sec (user-specified in front panel) before gathering data.
The program works, but I figured there had to be a better way. The thing about the timing, however, is that it changes from run to run. Sometimes a delay of 0.45 s works well, and other times the power supply will have shut down while the detector is still gathering data (and thus we miss the low end of the sweep). Other times the detector will turn on too early, which will cut off the high end of the sweep.
(Note: I have next to zero experience with LabVIEW, but I know a little bit of java, so I understand most programming jargon)
I spent all day yesterday trying to find out how to synchronize two data aquisitions (my attempt is shown in the "LIV SweepSummerDuncanRev3.vi" in the link at the bottom, 2nd image) but I'm running into trouble when I try to trigger the sample clock using the VISA Resource Name for the power supply.
The programs can be viewed here:
http://imgur.com/a/Up3eS
I'd really appreciate any and all advice that you folks have to offer!

Change the code such that rather than using a ramp from your power supply, just output a single value. Then do your measurement and then step to the next value in your ramp.
You can use a State Machine (SEarch if you do not know that term in LV).
Some gear will allow specifying a ramp driven by an external clock. If you widget can handle an external clock/trigger that approach could run faster.
Without hardware to syncronize the various sub-systems you would have to resort to using a RT OS and depending on the instruments, even that could be hit-n-miss.
So re-code to step-measure-step-measure etc.
Ben
Ben Rayner
I am currently active on.. MainStream Preppers
Rayner's Ridge is under construction

Similar Messages

  • Sharing VISA Resources With Multiple Application Instances

    I have noticed a strange behavior when attempting to communicate with a
    slow, single shared VISA resource (GPIB device) from multiple
    stand-alone LabView 5 applications.
    Initially, I developed some code which first attempts to acquire
    exclusive access to the device using VISA lock with a reasonably large
    timeout. After the lock is acquired, the code performs its IO, then
    releases the lock.
    I then wrote some wrapper code which basically tries to enter the (lock,
    IO, unlock) code as fast as possible. I built a stand-alone application,
    and ran 3 seperate instances. Basically, I could see that each
    instance would eventually acquire a lock, perform its IO, and unlock.
    However, some of the instances "starved" for access to the device,
    enough so that the timeout was sometime reached (each IO would take
    perhaps 2 seconds, and I set the lock timeout to say 10 seconds).
    Ideally, each instance should receive the lock in the order by which the
    lock was requested. It appears that VISA is doing something else.
    I've verified that this issue also exists with "faster" GPIB devices
    (say 200ms IO turnaround), but is less noticable because the
    communications are faster.
    To get around this, I wrote another application which basically acts as
    a resource server. Using VI server, each instance which wants access to
    the resource sends a "lock" request to the server. Specifically, the
    "client" calls a re-entrant "Interface" VI via VI Server. The
    "interface" VI then adds the "lock" request to the queue, and then waits
    for "Notification" from the server process. The server parses the
    request queue, and when the resource is available, it sends a
    notification to the appropriate "interface" VI. This whole time, the
    client is waiting for the interface VI to return, just like when it's
    waiting for a normal lock. When the client is done with the device, it
    sends an "unlock" message to the interface VI. The server then gives
    the resource to the next item in the queue.
    To me, this is a bit kludgey. My solution requires that this "server"
    application is running at all times, and it is a bit slower than the
    usual manner. Has anyone come up with a better solution?
    Thanks,
    Andy

    My answer depends how stand alone your applications are. I don't know what happens if you have several independant executable files but the simplest solution in the development environment is to have one vi that handles all send/receive requests to the instrument. In this way, without extra overhead, you guarantee single parts of each program access too your instrument. (don't make this vi reentrant naturally)
    This approach has another advantage. You separate in a better way the instrument and your aplication, so changing to another instrument is easier to do.
    greetings from the Netherlands

  • Synchroniz​ing two counter frequency inputs with multiple analog inputs

    Hello all,
    I'm fairly new to LabVIEW and I'm trying to collec​t data from multiple sources with synchronized tim​ing on the acquisition but I'm having trouble figu​ring it out. My problem is that I've got two count​er frequency inputs, one optical tachometer readin​g one pulse per revolution, and a max machinery fl​ow meter with a k factor of 12000. I can't seem to​ figure out how to sync the timing with my multiple analog inputs. I've be​en attempting to get the tachometer  to sync with ​the analog inputs first by following the example l​inked here. (https://decibel.ni.com/content/docs/DOC-10785) So far each time I run it I either get a timeout e​rror on the DAQmx read or a "Multiple sample clock​ pulses were detected" error (see attached image).  It seems if I slow the sampling rate way down to ​say 10 hz and ensure that the tachometer signal is​ over 800-1000 RPM (13-17 Hz) before starting the VI then the program will run without errors until ​the RPM drops below that threshold then the "Multi​ple sample clock pulses" error occurs.  The code is attached below.
    Does anyone know of a more effective way of syncin​g counter frequency inputs with analog inputs?  I'd like to have a VI that can show 0 RPM (and ev​entually 0 flow as well, but I think I need to fig​ure out the timing of one counter before I add ano​ther as it seems I can't have two counters in the ​same task). Any help on this would be greatly appr​eciated.
    LabVIEW version 13.0
    cDAQ-9178 Chassis with NI 9401 for the two counter inputs and NI 9205 for the analog inputs.
    Thanks!
    Richard
    Solved!
    Go to Solution.
    Attachments:
    SimpleDAQ.vi ‏44 KB
    LV_Error.JPG ‏31 KB

    Maybe third times the charm? 
    So I've finally got a good handle on why the VI is having problems at low RPM though I'm somewhat embarassed how long it took me to do that
    Because I have the counter time synced to my Analog input task if it doesn't see at least two pulses between the two clock pulses set by the analog input task I get the -201314 "Multiple sample clock pulses" error. This seems fine at first as it just sets a minimum RPM that I can measure and it's well below the area I'm interested in so no problems there.  I tried a simple error handler that would clear the error when it happend assuming the loop would keep iterating until the RPM went above that minimum at which point I would get a signal again. This is not the case, the read function just continues to spit out the -201314 error even after the RPM is back in the readable range. So then I tried adding two case structures so that when the error occured it would stop the task, clear the error, and then start the task again on the next loop iteration (Code Attached). This also doesn't work as the error shows up again on the stop task and then AGAIN on the start task on the next loop iteration. It seems this error is not actually being cleared and once it happens it stays with the task regardless of what the error cluster is carrying. 
    Anyone have any ideas?  The only solution I can think of is to just clear all tasks and recreate them each loop iteration until the RPM is readable again but that strikes me as a horribly clunky solution.
    Richard 
    Attachments:
    SimpleDAQ_1_Start Stop.vi ‏48 KB

  • Synchroniz​ing Torch 9800 contact list with my Playbook

    I am trying to avoid retyping all my contacts on my Torch 9800 into the Contact on my Playbook.  Is there a way to port them from my phone to the Playbook and then sychronize them so when one changes, it automatically changes the other?

    Hey pratiktnm,
    Welcome to the BlackBerry Support Community Forums.
    Thanks for the question.
    Are you setup on a BlackBerry Enterprise Server?  With the Desktop service book appearing it seems you were on one.  If you are still on a BlackBerry Enterprise Server, I would suggest creating a full backup with BlackBerry Desktop Software: www.blackberry.com/btsc/KB33505
    Then proceed with a security wipe by going to Options>Security>Security Wipe and then contact your network service provider to generate a new Enterprise Activation password.  Then proceed with the enterprise activation by going to Options>Advanced Options>Enterprise Activation.
    Once the activation is completed, test out the wireless synchronization of the contacts.
    Let me know if you have any more questions.
    Cheers.
    -ViciousFerret
    Come follow your BlackBerry Technical Team on Twitter! @BlackBerryHelp
    Be sure to click Like! for those who have helped you.
    Click  Accept as Solution for posts that have solved your issue(s)!

  • Synchroniz​ing - Data Recovery (.cvf) Help

    Newbie to the forum.  Hopefully this is the correct thread? 
    Had Palm Desktop  v 4.2 installed on my Windows XP Pro and was using it to sync with my Treo 650.  Recently purchased a Pre and am in the process of moving the data over.  Painful so far.  Palm support web site indicated that Palm Desktop v6.2 was required to begin the data tranfer to the Pre.  Downloaded v6.2 and installed.  The install recognized the previous version 4.2. and the install seemed to go well.  When I opened the new Desktop none of my old data (contacts, etc.) was there.  I was especially confused that no contact catagory was shown with the new install. 
    Prior to doing this upgrade, I did export all of the data to a .cvf file on my desktop.  I believe I was asked and did sync the Treo with the new v6.2 as part of the install..  Because none of the data was transferred, I tried to import the saved .cvf data, but the new Palm Desktop version only recognized a new .cvs format . . .   And all of the data is still available on my Treo . . .
    Frustrated and thinking I should start over I did uninstall the Palm v6.2 completely (Windows add/remove).  Suggestions . . . ?
    Post relates to: Pre p100eww (Sprint)

    A little confusing... There are two common export formats for contacts (address book entries):
    .vcf = vCard format, most often used for individual export
    .csv = comma separated values, most often used for bulk export
    If your address book on your 650 is intact, you may be able to sync the data into Desktop 6.2. However, I haven't used that version myself. You might peruse the Windows HotSync board here for pointers on the 4.x to 6.x migration.
    http://forums.palm.com/palm/board?board.id=windows​_hotsync

  • Synchroniz​ing Palm Treo 680 Unlocked with Outlook 2003 on Windows 7

    After 2 full install and reinstall of Palm 6.2 I cannot synch my Outlook 2003 and Palm Treo 680
    Hotsynch Journal says everything has synchronized correctly but new meetings and task changes in one platform do not appear on other platform.
    There is an error mentioned for file PmTrcae Database (Protocol error 4004)
    I must mention that Installation process seems to drop after announcing that first hotsynch will be performed, it never does, I had to start it manually from the Treo.
    Will there be a proper Windows 7 version published?
    Post relates to: Treo 680 (Unlocked GSM)
    This question was solved.
    View Solution.

    gallinago wrote:
    I downloaded v.6 onto my new Windows 7 computer. When I try to hot sync, I get a window stating that the driver is missing and I cannot find a driver on your site or on windows to download.  I don't think Outlook is involved here; I don't use it.
    Are you running a 64bit version of Win7?
    As reported many, many times in this forum, there are no USB drivers for this version.
    Please read the Sticky post at the top of this section for guidance:
    http://forums.palm.com/t5/Windows-HotSync/Windows-​7-and-Vista-hotsync-instructions-here/td-p/251755/​...
    WyreNut
    I am a Volunteer here, not employed by HP.
    You too can become an HP Expert! Details HERE!
    If my post has helped you, click the Kudos Thumbs up!
    If it solved your issue, Click the "Accept as Solution" button so others can benefit from the question you asked!

  • Extract device from physical channel

    As always I want to do something absolutely trivial and am running into all sorts of issues
    I define (through a control) a physical channel that I want to take data on. I want to be able to programmatically reset the device the channel is on. i.e. i don't want someone to have to open up the code and change the device on the reset vi if necessary. Even if there is only one device physically connected, you can’t guarantee the device will be Dev1, because Labview keeps a history of devices that have been plugged in. The obvious solution – extract the device name from the physical channel control which contains the channel(s) you have selcted. Can I find a trivial way to do it, NO!  There is a property node called device but it does not (despite its name) return the device you have selected along with the physical channel.  There are convoluted ways of doing this, but i'm hoping there is some easy way. Suggestions
    Solved!
    Go to Solution.

    I think this might be the "trivial" way you're looking for: some string pattern matching and a simple
    regular expression to extract the device name prefix from a physical channel name.
    -Kevin P

  • How can i synchronize the data from PDA (address book and calendar) with AP

    I�m yang J2ME programmer. I�m just programming an API. I need to synchronize the data from PDA or mobile phone (address book and calendar) with PCs .
    Is the JSR 75 s right way? or JSR 230?
    Where can i download JSR 230 ? not PDF

    Can somebody help me?
    I mean, i want to read address book from cell phone to PC.
    Message was edited by:
    cash_ye

  • How use USB with VISA resource name?

    Hi, 
    I'm new at LabVIEW and I'm working on a code that has to acquire data from a measurement device.This device only has a USB output so I have to work with it. As far as know the block that is used to refer a external port is VISA resource name. I have installed a driver that recognizes the serial ports, and they appear in the list of VISA resource file at front panel. So I figured that installing the product's USB driver so the USB's ports would appear. I installed and nothing of USB ports. I don't know it is that simple or working with USB ports is a little bit harder. I've tried a lot of things and nothing worked so here I am.
    PS: I have noticed that the driver's version is 64 bits (my pc is 64 bits)  and my LabVIEW version is 32 bits, so could that be the problem?
    The device by the way is Digital Analyzer WT500, Yokogawa.
    Thank you.
    Solved!
    Go to Solution.

    Victor,
    the suggestion of Instrument Driver is to make more easy using the instrument. 
    Normally, you should install a windows drive from manufacturer (if it isn't installed, the instrument will appear like a pendrive in device manager) and the NI VISA. But for some instruments, the manufacturers have API's that uses NI VISA to communicate with your equipments.
    So check if your instrument appear correctly in device manager (as indicated by Yokogawa). If not, check if Yokogawa has an appropriated driver for this instrument and install it. 
    Is it clear?
    If you prefer, contact NI Brazil Support by website or by phone (11)3149-3149 Ext 3.
    Ricardo Ramos
    Engenheiro de Vendas - Sul
    National Instruments Brasil

  • How pass VISA resource string from C++ as uintptr_t* to a LabVIEW generated DLL ?

    My coworker build a DLL from LabVIEW that I am trying to call from C++.
    void__cdeclMeasureDCvolts(uintptr_t*VISAResourceName, int32_tChannelNumber,LVBoolean*MonitorResult,double*MeasuredVoltage, TD1*errorOut);
    Does "extcode.h" define macros or functions for string conversions?  Or should I request my coworker rebuild the VI's to make arguments more C/C++ friendly?
    Thanks in advance for any tips or direction.  Examples are * much * appreciated.
    -Ed
    TestObjectMeasurement* testObjectMeasurement = m_measurementMap[name];
    TestObjectInstrument* testObjectInstrument = m_instrumentMap[selectedInstrumentName];
    int slot = 1;
    int bank = 1;
    int channel = 1;
    int channelNumber = 1000 * slot + channel;
    char * visaResourceName = testObjectInstrument->visaResourceName().toLatin1().data();
    LVBoolean monitorResult = LVBooleanFalse;
    double measuredVoltage = -1.0;
    TD1 errorOut = {0};
    try
    // void __cdecl MeasureDCvolts(uintptr_t *VISAResourceName,
    // int32_t ChannelNumber, LVBoolean *MonitorResult, double *MeasuredVoltage,
    // TD1 *errorOut);
    MeasureDCvolts(reinterpret_cast<uintptr_t*>(visaResourceName), channelNumber, &monitorResult, &measuredVoltage, &errorOut);
    catch(...)
    qDebug() << "Exception thrown by: MeasureDCvolts";
    QString errorSource = QString::fromUtf8(reinterpret_cast<const char*>(LHStrBuf(errorOut.source)), LHStrLen(errorOut.source));
    qDebug() << errorSource;
    LStrHandle errorTextHandle = {0};
    bool foundErrorText = NIGetOneErrorCode(errorOut.code, &errorTextHandle);
    if(foundErrorText)
    QString errorText = QString::fromUtf8(reinterpret_cast<const char*>(LHStrBuf(errorTextHandle)), LHStrLen(errorTextHandle));
    MessageHelper::messageError(errorText, errorSource);

    Why is it not a solution? It was exactly about what you wanted to do and the solution in the last post is how it can be made to work. There is NO way to ddirectly create a LabVIEW native VISA resource in C code that is not called from within LabVIEW, unless you consider hacking the LabVIEW system and using undocumented functions a solution. But that has very good chances of breaking with future LabVIEW versions.
    Your only other option is to export another function from the DLL that accepts a String and uses VISA Open to explicitedly open the VISA session in your LabVIEW DLL and returning that VISA session to your C code. This is functionally equivalent to passing a String to your current function but avoids the lookup and potential recreation of the VISA refnum on each call. Something I would only feel bothered if you intend to call this function many thousend times each second, which for VISA communication seems very unlikely.
    Rolf Kalbermatter
    CIT Engineering Netherlands
    a division of Test & Measurement Solutions

  • Working with VISA resources in Labview 6

    I am having difficulties programming through VISA in Labview 6.0.2. I have installed MAX 3.2, and have configured my two devices going through USB:
    1) an Agilent 82357A USBtoGPIB adapter connected with an Agilent 53131A Universal Counter
    2) a USBtoRS422 converter registered as serial COM4 connected with a Rainin Peristaltic pump
    I can open communication with both devices from within MAX, and send/receive data with no problem.
    When I try to program in Labview 6, I cannot communicate with either VISA device, even though I am using VISA instrument IO vi's with the configured VISA resource name from MAX. I'm not sure what I'm doing wrong.
    Is there an incompatability between MAX 3.2 and Labview 6.0.2, or have I configured something wrong?
    Thank you for the assistance.

    Hello,
    I think we can defeat the problem with Device 1 fairly quickly. If you have established that you can communicate with the device in MAX, I think we simply have a small programmatic change to make in LabVIEW. I am interested in having you attempt using the VI I have attached, which allows you to specify the bytes you would like to send in an array, and then simply sends the corresponding the string of ascii characters (the characters corresponding to the values in the array); this way we can be sure to send exactly what we would like. The following link has an ascii table in it, which you can use as a reference when putting the values into the array: (where the first byte specified will be the first character in the string)
    http://www.ni.com/pdf/manuals/371253a.pdf
    Simply change the VISA resource name to reflect your device and give it a try.
    As for Device 2, let's try using the same VI (attached) and write a command which the device is known to understand. A simply query would be approprite, where you can write using the attached VI, and then read using another VI which literally only has a VISA Read function in it (and a resource control or constant connected with your resource specified).
    One thing to verify before reposting is whether the devices you have use EOS (end of string) and/or EOI (end or identify) termination of GPIB operations. For EOS, a character is sought by the listener to indicate the end of a transmitted string, such as a linefeed or carriage return (or both). EOI termination is defined by the talker asserting the EOI line when the last byte to be transmitted is sent; the listener knows it has received the last byte of that transmission.
    Definitely repost; it is likely that we are only few programmatic changes and perhaps a spec check (for the devices) away from dependable communication
    Thank you,
    JLS
    Best,
    JLS
    Sixclear
    Attachments:
    WriteByteArray.vi ‏14 KB

  • Old visa open, error code 1073807343, using VISA with a GPIB device, and VISA resource names

    Hi everyone,
    I'm trying to get a SRS model SR720 LCR meter (manual on this page) working with LabView, connecting it to my PC with an Agilent 82357A GPIB/USB interface (manual on this page). I'm running MAX version 14.0, and NI-VISA version 14.0. I'm running LabView version 13.0f2, 32 bit.
    I followed the directions on this page, and I'm pretty confident I can communicate with my LCR meter, because I can go to NI MAX->GPIB0::17::INSTR under "devices and interfaces", then go to the VISA test panel, go to the I/O tab, and I can enter a few commands from the SR720 manual, like changing the frequency, and I see the light on the LCR meter change to what I just told it to do. Here is a picture, just to illustrate it:
    Also, if I do the *IDN? command, it returns "
    9: Write Operation (*IDN?)
    Return Count: 5 bytes
    10: Read Operation
    Return Count: 41 bytes
    StanfordResearchSystems,SR720,08087,1.03\n
     So, that's good too. I also made the alias of this device "MyLCRmeter", but that shouldn't matter for now.
    Now I'm trying to get it to work in a more usable way. The SR7xx drivers that LabView found for me came with a couple sample programs, one called "Getting Started.vi". I tried to just simply see if it worked, changing the "instrument descriptor" field to "GPIB0::17::INSTR", since that's the one that seemed to be the right device in MAX. Here is a picture of the front panel of "Getting Started.vi":
    Trying to run it like this gave me this error (code -1073807343):
    Old VISA Open in SR715/720 Initialize.vi->SR715/720 Getting Started.vi
    Doing a bit of searching on these forums, I found these few threads (1 2 3), which seem to tell me that, since this program was written, the way you open VISA has changed. In the 2nd of those links, one guy said:
    You just have to replace the Old VISA Open function with the VISA Open that is on the Instrument I/O>VISA>VISA Advanced palette. When you do that, the wire to the string that was used for the resource name will be broken. Delete the string and make the actual VISA Resource Name visible on the front panel. Wire the VISA Resource Name control to the connector pane where the string control was (upper left connection).
    So, I went into the sub-VI Initialize.vi, and did that. Here is a picture of what the front panel and block diagram of the sub-VI Initialize2.vi (I changed the name in case I screwed things up, but I'm pretty sure the new one is the one being called by Getting Started.vi) looked like before I changed anything (you can see the Old VISA Open):
    and here it is after replacing that with the regular VISA Open, deleting the instrument descriptor string and replacing it with a VISA resource, and connecting that to the VISA Open:
    However, as you can see in the image above, and it seems like someone else had this problem in the 2nd thread linked above, I don't see my device listed in the drop-down menu (even if I refresh), only "LPT1" (I don't even know what that is):
    No, I wasnt able to select the visa resource from the pull down menu. somehow it seems disabled or something.
     If I just enter the VISA resource name from MAX manually, GPIB0::17::INSTR, and save and try running that, I get the error(code -1073807346):
    VISA Open in SR715/720 Initialize2.vi->SR715/720 Getting Started.vi
    So I'm not sure where to go from there, or why it's not in the drop down menu to begin with. I'm sorry if I've missed something but I've searched about as far as I can go. If I had to guess, maybe one thing could be that in MAX, under Tools->NI-VISA->VISA Options->My System->Conflict manager, I have both Agilent VISA and NI VISA enabled; I know sometimes problems arise from conflicts between different software fighting over the same hardware. Is this okay, or do I need to change it?
    Thank you and please let me know if there is any other code/screenshots I could provide that could help!

    Hi, I did indeed enable NiVisaTulip.dll under MAX->Tools->NI-VISA->VISA Options->My system->Passports->List of passports. Sorry, I should have mentioned that explicitly.
    By install as primary, do you just mean as it is here, where it is selected as the "Preferred VISA"?
    Or somewhere else? I installed NI-VISA first I believe, and then the Agilent one, and during the installation of the Agilent one, I'm pretty sure I made it not the primary. Here's what I have in the Agilent (it's called Keysight now) Connection Expert software:
    I also have this under the "Keysight 488 options" tab:
    Should that be checked? I don't actually know the role of VISA vs 488 here...
    What could I try?
    Thank you!

  • Using VISA Resource control with Windows 7 64-bit

    Hi.
    My company recently upgraded our laptops, and I'm now working on Windows 7 64-bit rather than Windows XP 32-bit.  I reinstalled LabVIEW 2009 Base Package, trusting that the Windows 7 64-bit OS would be able to handle LabVIEW 2009 32-bit without problems.  But when trying to run our test system, the VISA Resource control doesn't work.
    We use it to get a list of the available COM (RS232) ports on the computer.  But now the list comes up blank.  The only option that comes up in the list, is the text "Refresh".  But nothing happens when I press it.  Is this a bug in LabVIEW?  I've installed the latest patch I found, so now I'm running LabVIEW v9.0f3 (32-bit).
    Thanks alot for any and all assistance!
    Solved!
    Go to Solution.

    Sorry for not replying to you earlier.  I've been busy on other projects.
    I'm not sure what you refer to with MAX, but I'm able to access the com ports using other tools/loggers.  So it's not an issue with Windows 7 itself.  But it seems logical that the problem would have something to do with running 32-bit drivers on a 64-bit OS.  Referring to this list from NI (http://digital.ni.com/public.nsf/allkb/71E9408E6DEAD76C8625760B006B6F98), I'm downloading the 64-bit installer now.
    But if I understand it correctly, you cannot run applications from a 64-bit Application Builder on a 32-bit OS.  So basically you need the Application Builder installed on both a 32-bit and 64-bit OS to have all bases covered.  But in the FAQ, it's asked if it's possible to have both 32-bit and 64-bit LabVIEW 2009 installed on the same 64-bit OS.  Why would this be wanted?  If you cannot use the 32-bit Application Builder, and you cannot trust that the 32-bit LabVIEW runs successfully (i.e. driver issues etc.), why would you want to have a 32-bit installation on a 64-bit OS?

  • Missing TCP/IP Socket from VISA Resource Property

    My Labview "VISA Resource Property" is not showing "TCP/IP Socket".  According to this website, it should:
    http://zone.ni.com/reference/en-XX/help/371361D-01/lvvisaprop/visa_object/
    I/O Session
    FireWire Instr
    GPIB BoardInterface
    GPIB Instr
    Instr
    PXI Backplane
    PXI Instr
    PXI MemAcc
    Serial Instr
    TCP/IP Instr
    TCP/IP Socket  RIGHT HERE
    USB Instr
    USB Raw
    VXI Servant
    VXI/GPIB-VXI Backplane
    VXI/GPIB-VXI MBD Instr
    VXI/GPIB-VXI/VME MemAcc
    VXI/GPIB-VXI/VME RBD Instr
    I'm using Labview 8.6.1 along with MAX 4.6.x  I'm able to see & utilize VISA TCP/IP Socket in MAX so I'm wondering how come I don't see it in Labview?  Is there a way to enable this?

    nevermind, I found a workaround.  I just type cast "TCP/IP Socket" from a "instr" VISA control

  • OSS functionality and synchronization of data with SAP SOLMAN 4.0

    Hi experts,
    I would like to enable the OSS functionality in our SAP SOLMAN 4.0.
    I need a document for the settings to activate the OSS functionality.
    I am looking forward to hearing from you,
    Thanks alot.
    Tolga AKINCI

    If your system has more than one IP/multiple network cards, take care, that not the wrong addresse is sent to OSS.
    We have had quite big trouble last month, when our configuration in OSS got overwritten by wrong IPs, the system was unaccessible and thus support could not access our systems and eventually payroll run one day later and ~ 600 people were angry.
    Our connection was working the last 10 years so nobody came to the idea, that SolMan is the root cause. Since there´s no option to choose, which IP is finally sent to OSS but only a "manual coding" (note 1050693 - Several IP addresses when you synchronize system data) we decided finally to delete all systems/solutions from SMSY.
    So be VERY careful what you synchronize.
    @Rüdiger: I appreciate your enthusiasm, but I got hit just too often with that SolMan things, spent hours and hours because SAPs "labour" environments are clean and tidy. I can just warn people out of my experiences to use too much "automatisms" before any harm is done (and yes, we got substantial harm, not only virtually and in the mind but also $$$).
    Markus

Maybe you are looking for

  • What app do I need to install on my iphone so that audio files that have been shared with me will play?

    I have google drive on my phone but I can not play any audio files that are shared with me because it keeps telling me that it is not supported without the right app.

  • View photos and videos separately

    Hello Is there any way to view photos and vides separately on IOS 7. I find so annoying looking for a video in the middle of so many photos. On iOS 6 it was easy but I can't find how to do it on iOS 7. Thanks!

  • Static coming through computer line out, PLEK500

    I have three of these working just fine, in this configuration: #1 tied to cable modem #2 tied to ethernet port of a macbook pro #3 tied to a Roku box #4 is meant to serve the Ethernet port on the Mac Pro tower in my home office.  Unfortunately, upon

  • No more wifi detection,

    After last software update yesterday, the ATV3 would block after language question and keep repeating that screen. Tried to manually upgrade with mini USB and iTunes but no change, it stayed in a loop with that question after boot. So I installed pre

  • ME51N Change G/L Account default for Material group.

    Hi experts, I am trying to create Purchase requisitions(ME51N) for Account assignment catergory K-Cost centre. When i am entering the material group its populate G/L account as default in account assignment TAB. how to change this default. What is th