LabVIEW since 3.0 and TestStand since 1.0

Highly experienced in all types of instrument communication and NI DAQ. I'm looking for full or part time work in the Boulder, CO area. PM me for details and resume.
Thanks

Hi Faisal,
PI 7.1 is more advanced than XI 3.0:
PI 7.1 Features:
A new Enterprise Services Repository which can be used both at design time and run time. The repository is "high volume ready" and supports UDDI v3.0. The repository implements governance processes and offer a central visibility of the services and their artifacts. The repository is of course loaded with SAP’s enterprise services classification and enterprise services definitions.
Full support for the latest web services standards including WS-Policy, WS-TX and support for WS-ReliableMessaging (WS-RM).
Support for incoming message validation using XML Schema
Support for Events which are a key element of the product. It introduces the notion of Global Events available across all processes to be reacted upon.
Message bulking for asynchronous invocations has been introduced which helps throughput by a factor of 3 to 4.
Support for principal propagation which is a pre-requisite for building composite applications.
NWPI v7.1 is built on a JEE 5 foundation which helps memory consumption. However, it appears that NWPI does not support WS-BPEL (other than as an import format), BPEL4PEOPLE, SCA or SDO just yet.
Folders are introduced with the development tools in NW PI 7.1. Learn the pros and cons of using folders in PI 7.1 to organize projects and interfaces. XI 3.0 is the XI protocol valid for both SAP NetWeaver ´04 and SAP NetWeaver 7.0.
SAP® Exchange Infrastructure 3.0: Demo Examples
https://www.sdn.sap.com/irj/servlet/prt/portal/prtroot/docs/library/uuid/d46b9b90-0201-0010-ef92-9964c1bc98e1
Implementing a Distributed Two-Phase-Commit Scenario with Web Services and SAP NetWeaver PI 7.1
https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/00ffdb4d-e869-2a10-7688-891d7eea1b12
Preview of Web Services Reliable Messaging in SAP NetWeaver Process Integration 7.1
https://www.sdn.sap.com/irj/sdn/go/portal/prtroot/docs/library/uuid/304335f7-f33c-2a10-ae80-9c9ffdc86415
Regards,
Vinod.

Similar Messages

  • Controls System Engineer-Labview and Teststand

    One year experience in Labview and Teststand/automated testing system for Locomotives.Masters in Embedded Software. Seeking Employment Fulltime/Contract.

    Hi,
    Thank you for responding. I am based at USA. Currently looking for opportunities here.
    Thanks!
    Priya

  • Labview edit time menu and right click popup menus

    How can i add a menu item to the labview edit time menus and to the right click popup menu of any selected object in a block diagram?
    I would like to integrate commonly used tools by me out of the springboard scripting tool into actual labview.
    Solved!
    Go to Solution.

    It sounds to me like you are looking for something like the Right-Click Framework.  Some people really like it, I have not used it extensively myself.  I've had a request to convert a quick drop shortcut into a RCF plug-in, I'll see how it goes.  I have also longed for a good way to hook into the LV IDE to add my own tools, what I find are useful, but not fully satisfactory solutions.
    1)  Right-Click Framework:  Some initial startup costs in terms of downloading and installing.  Start up is a bit slow (on my machine), and the workflow is a bit different than what I am used to:  Select an object, press the hotkey, right-click.  You can get used to it, but it is not as responsive as a built-in right-click since you can't right-click in one action.  It seems to be easily extensible, I'll know more in a day or two.
    2)  Quick Drop:  Built into LV.  Slow startup like RCF.  Fairly easily extensible.  Not context aware (unless you build it in to your plug-in), so you have to choose your letters wisely.  After the initial pain, the response time is not shabby, does require the extra Ctrl-Space action.  
    3)  Tools Menu:  I drop a few things in here, like my custom connector pane editor.  Not bad, but there are no keyboard shortcuts (what a pain), so it is good for actions that you might call once or twice on a given VI.
    In short, all of these methods require extra clicks or keypresses and are a bit sluggish.  If something like copy and paste behaved this way we would never use LV again, but if your tool truly saves you a bunch of steps it is well worth it.  I have almost run out of letters to use for my QD shortcuts.
    Addedendum:  Personally I am not quite ready to start talking to my computer, but LVSpeak seems pretty cool, Smashy Smashy.

  • How fast does Labview perform integer, float and matrix calculations?

    What is the speed of a modern CPU? Ok it is 3Ghz, one to four cores etc. but how much work can I get done in one clock cycle at this speed?
    This kind of questions have been in the back of my head for some time, so I decided to test it. Of course using Labview since it so easy to to use!
    I wrote a really simple program that performed 10000 adds 3 to an 32 bit integer in a loop. It then repeats this until 1000ms has elapsed and calculates the MegaInts (or whatever I should call it). I then did the same thing by adding 3 to a matrix of 10000 32bit ints.
    On and dual core 3Ghz PC from 2005 I got 261 Mints in a loop, and 1400 Mints in matrix. Only one core seems to be used, CPU load aorund 50%. So every two clock cycles or so I get some work done in a matrix calculation. Also older 1,8Ghz Celeron computers performs about the same in this test, even though they feel very slow in normal use.
    On a newer 4/8 core 2,8Ghz PC from this year I got 830 Mints in a loop, and 5900 Mints in matrix. Only one core seems to be used, CPU load aorund 12%. So actually more than one add per clock cycle!
    This indicates that either the compiler is smarter than me and does not to all integer adds, or the CPU is smart and do them very fast!
    See this as not the complete benchmark of Labview in integer, float and matrix calculations, but as a teaser to make you do your own tests.
    Play around and have fun!

    Norbert_B wrote:
    Yamaeda wrote:
    A standard loop only runs in 1 thread,[...]
    That is not true.
    A standard loop follows the clumping algorithm during compilation. Clumps can be distributed on any number of threads.
    So, a standard loop CAN be distributed between several threads.
    There are, however, some settings and structures which will result in a single thread per loop:
    - Setting the calling VI to "subroutine" priority (not recommended)
    - Using a Timed Loop instead of a standard one
    - AFAIK: Containing the loop in an InPlace Element Structure with "Data Value Reference" border nodes
    There is at least one additional option, but it is so unlikely (as it messes up with most of LV!) that i will not point it out here.....
    Norbert
    Why would we have parallellization options for a loop if it's already multithreaded? I'd love if loops would automatically parallellize.
    If you mean the content inside the loop, then it'll follow ordinary optimization and use several threads if possible. In this case OP mentioned 1 cpu was at 100%.
    /Y
    LabVIEW 8.2 - 2014
    "Only dead fish swim downstream" - "My life for Kudos!" - "Dumb people repeat old mistakes - smart ones create new ones."
    G# - Free award winning reference based OOP for LV

  • Labview DSC 8.5 and Lookout 6.1

    Ok...so I'm moving to some new hardware and new OS's along with Labview upgrade and Lk 6.1.
    I have not separated these stages yet so I'm hoping NI can help me sort this out.
    The scenario:
    Run Windows Server 2008 and install LabVIEW 8.5, then LabVIEW DSC 8.5, then Lookout 6.1
    Lookout hangs on startup!
    Now windows Vista with SP1 is essentially the same as Win2k8 with SP1, so the question(s) is/are....
    1.)  Is Lookout 6.1 compatible with Vista SP1?
    2.)  Will Lookout 6.1 run properly installed on top of LabVIEW 8.5 w/ DSC?  and/or Can LabVIEW DSC 8.5 and Lookout 6.1 coexist and work properly?  If yes, can they run simultaneously on the same machine?
    3.) Does Vista SP1 break Lookout 6.1?  If not, why does it break on W2k8 server?
    I understand NI does not support server OS's  (can't really figure that out since Lookout actually runs more stable with a server OS....) but Lookout has always run fine on Windows Server OS's since NT4.
    I just want to know whether to bother setting up the same software scenario on Vista w/o SP1, or if you NI wizards know for sure that something else is breaking the scenario...
    Thanks,
    Ed

    1.  We can't say whether Lookout 6.1 supports Vista sp1 or not. Officially saying lookout 6.1 does support Vista, but sp1 is newly released and we haven't tested yet, so if it runs fine on sp1, it's good, but we can't guarantee. We will test the lastest os and sp for the next release of lookout.
    2. Yes. You can install both on the same machine and run both, at least on the supported os.
    3.  We haven't tested it yet.
    We don't support server os is because we have never tested the server os. But it doesn't mean lookout can't run on server os. Maybe it runs fine, maybe not. So I'm not suprised that Lookout has always run fine on Windows Server OS's since NT4.
    For Vista sp1, I'm sorry that you need to try by yourself.
    For the SPC toolkit for DSC 8.5, it's a known issue. It has been fixed in DSC 8.5.1.
    Ryan Shi
    National Instruments

  • LabView fpga VHDL code and compiler

    Hello,
    I'm in the project where we would like to use NI hardware (more likely cRIO system). With NI hardware we will read/wright several AI/AO and DIO and perform some math and controls on the result of readings. We are planning to design FPGA code for project, but we are thinking about implement all data processing and control logic in VHDL and link it with AI, AO and DIO with help CLIP or IP Integration Node as explained in this : "white-paper": http://www.ni.com/white-paper/7444/en/
    Mentioned above paper explain how to implement VHDL code in LabVIEW FPGA VI using CLIP or IP Integration Node, but the topic that is not highlight explicitly is how these construction CLIP and IP Integration Node will be handled by Compiler. The main reason for such approach (VHDL linked with part that read/write into hardware AI AO and DIO) we expect that our VHDL code will be handled by LabVIEW compiler without modification and passed to Xilinx Compiler synthesis as is (path for Compile process I've taken from here: http://www.ni.com/white-paper/9381/en/ ), so we will be able at some level bypass the intermediate process of compilation and get almost the same result as if we design pure VHDL code and use Xilinx ISE for Synthesis Mapping and Bit File generation.
    Will this approach work? I was not able to find any documents that explain the Compiler behavior and confirm that VHDL code handled untouched or will modified, does such document exist?
    Note. I've requested official  assistance from NI support on topic above, but I would like to post this question on forum hoping get more feedback.

    Hello RangerOne,
    There won't be any modications to the internal logic of the VHDL that you implement in the IP integration node. Though I've seen developers unfamiliar with LabVIEW FPGA get tripped up on the synchronization registers that LabVIEW FPGA inserts into the code around the integration node. Learning where and why these syncrhonization registers are inserted has in my experience always resolved this issue. These two help documents do a good job of explaining the 'where and why' of synch registers when the enable chain is present, or when working with IO inside of a SCTL.  
    With regards to the stability of LabVIEW FPGA, I would second Daniel's sentiments. What about the known issues list conveys instability and risk? As a point of comparison, here are the known issues for ISE 14.x. 
    If you are looking to minimize risk, I would recommend developing the critical logic in the development enviroment in which you are comfortable setting up a comprehensive test bench since testing the code is the only way to truly verify its functionality. For me this would be LabVIEW FPGA as it has excellent trouble shooting tools and I've been developing in it for quite some time. Perhaps you're more familiar with ISE than LabVIEW FPGA and that is the source of your trepidation? If that is the case then you may find the High Performance FPGA Developers Guide a good read.  You may also find a few of the case studies on our website reassuring since they demonstrate other teams successfully implementing a solution using LabVIEW FPGA. Here's one that used LabVIEW FPGA in conjnction with VHDL IP similiar to what you are doing.
    National Instruments
    FlexRIO Product Support Engineer

  • LabVIEW Run Time Engine and ActiveX Automation Server

    Hi,
    I have a question related to LabVIEW Run Time Engine and ActiveX Automation Server:
    We have a tool developed in LabVIEW and TestStand. End user interface is a LabVIEW GUI (it's an executable) which executes TestStand sequences. Some TestStand sequences call lower level LabVIEW VIs.
    Everything works fine with LabVIEW installed, but I was wondering if I could run this tool on deployment machine with only LabVIEW run time engine installed. We are using LabVIEW 6.1 and with only the run time engine installed, I get the error message: "Unable to launch the 'LabVIEW.Application' ActiveX automation server ". 
    BTW, we are planning to upgrade the versions soon.
    Is it possible to make things work only with the run time engine?
    Thank you.

    Thanks for the response, Brian.
    I did upgrade to LabVIEW 8.0 and TestStand 3.5. Everything works fine on the development computer where LabVIEW and TestStand is installed. On the target computer in the lab, I installed LV RTE 8.0 and TestStand 3.5, and now when a VI is called in a sequence file, I get the following:
    RunState.PreviousStep.Result.Error.Code  -18002 
    RunState.PreviousStep.Result.Error.Msg  "LabVIEW: The VI is not executable"
    So, the executable opens and works fine (which means that the LV RTE is installed properly), but when a sequence calls a low level VI, it does not make it through.
    comments/suggestions are appreciated. Thanks.

  • Labview program to unistall and install new/updated software programatically

    Hi,
     I want to create a labview program to install and uninstall teststand deployment programatically.
    Query : I have teststand deployment made and ready,wanted to deploy in 5 test PC's and for very first time it did work good. Now when I made updates in the modules, i made a new deployment with version increment. Now this time, i first uninstalled the previous deployment and reinstalled a new one. This time it took time to uninstall and install all 5 test PC's.
    I wanted to automate this process through labview. I wanted to create a labview exe or installer which continusouly running on the client PC's looking for a perticular folder for new/updated deployment. If it finds, uninstall the older version if present and install a newer version of deployment.
    I wanted to know what things i need to consider while making this type of labview program which uninstalls the older version and installs a newer one? I think i need to even reboot after uninstallation, before installing newer deployment. How make the program running even after reboot and continue the process?

    I have the MS-6968 device and i want to use driver 1.4.x or 3.0.x. I found the driver 1.4.2.10 but i cannot find any WIDCOMMSecurity code.
    So, can anyone give me a link for BTWPatcher.exe???
    It cannot be found here (Jon's Bluetooth ) anymore.
    Thanks

  • Calling labview 8.2 Vi from teststand 3.5 throws an error

    calling labview-8.2 vi in teststand-3.5 throws the following error.
    In the development environment, where Teststand 3.5 and LabVIEW 8.2 are installed, teststand file doesnt throw any error. However when the same sequence is deployed in a machine that doesnt have teststand and labview. (Both Teststand runtime engine and Labview runtime engine) are installed in the end machine.
    Trying the execute the sequence in the end machine throws the following error
    "Error loading step 'Call LabVIEW measurement module for amplitude imbalance' of sequence 'MainSequence' in file 'TSG06-Amplitude Imbalance.seq'.
    LabVIEW:  VI version is later than the current LabVIEW version."
    Attachments:
    Error.JPG ‏26 KB

    Hi,
    Should have also asked this question before.
    What is your LabVIEW Adaptor set too on the target system. Use ActiveX Runtime Server or LabVIEW RTE. If its set to use the LaBVIEW ActiveX Runtime Server which can be either the Operator Interface or the seperate ActiveX Server, then these are probably built with an early labVIEW.
    If so you would need to rebuild the server you are using.
    Regards
    Ray Farmer
    Regards
    Ray Farmer

  • LV producer consumer and TestStand

    In the TestStand process model, I've replaced the original Get UUT Serial Number dialog with a custom dialog that has some additional fields, and will (eventually) allow the operator to BarCode scan the serial number.
    I'm thinking about using LabVIEW's template for Producer/Consumer.  I basically want to allow the user to hang out at this dialog until the click on either "START" or "CANCEL".  They can tab to the different fields, and Scan in the serial number if desired.  What types of 'events' would I want to capture?  Is this even the right design pattern for this situation?
    Also, this dialog needs to be modal (in front of TestStand executive). 
    Thoughts?

    mrBean,
     By custom dialog do you mean a seperate VI that you open and show the front panel when the step executes?
    If that is the case the producer consumer might be overkill. From what I've seen one of the main reasons one would use this pattern is to seperate the acquisition loop from the data record loop. In that manner the data recording doesn't slow down the data acquisition. I've used a wedge type barcode scanner in the past and it is pretty easy to impliment.
    On a side note I really enjoy reading your posts on both of the forum boards. I too at some later date will be creating a custom operator interface using LabView and your questions have been really helpful in my formulation of a plan. Thanks
    Using LabVIEW 2010SP1 and TestStand 4.5

  • LabVIEW: The VI is not executable. Most likely the VI is broken or one of its subVIs cannot be located. Open the VI in LabVIEW using File Open and verify that it is runnable.

    When trying to run a sequence on a deployment machine I get this error. When I switch the Adaptor back to the Development System the problem goes away however I do not have the luxury of keeping the full Labview development on the test machine. Is there a know fix for this problem and what is the cause of this error. Here is the software and versions I am using. It seems that the vi's that are broken are using Invoke and Property Nodes within them, I am not sure if this is just a coincidence. It appears it is more related to the invoke node than the property node. Any help is greatly appreciated. Why does it seem everything works great until you try and deploy it on a test machine. This is not the first time I have had these kinds of problems using Nation Instruments software.
    Labview 8.2.1
    TestStand 4.0
    Solved!
    Go to Solution.

    I have a similar problem.
    A bit of background - I had developed a test suite with a Labview front panel for manual operations, and a Teststand project which used some of the same lower level vi's, and specific test step vi's. It was developed on LV 2010 32-bit and Teststand 2010. All was fine.
    Recently i had to rebuild my development machine and installed only LV 64-bit 2011 and the latest (2010) Teststand from the 2012 Dev Suite.
    I knew that i wanted to run some new hardware drivers in 64-bit versions due to another problem, so i replaced the .dll in the very lowest level VI with the 64 bit version. And ran the manual interface. All ok.
    But when i try the Teststand execution, i get one of two errors:
    - If i set Teststand to use the LV runtime engine 11.0.1 (2011 SP1), the i get a "Unable to load VI xxxxx with the runtime engine version '11.0'" Despite the fact that each VI does run correctly in LV dev when loaded individually, and i've saved them as version 11.
    - If i set it to use the development system (which shows in the config menu 'active version:32-bit' even though 32-bit LV is not installed), then i get the "The VI is not executable. Most likely the VI is broken or one of its subVIs cannot be located." error, even though i when i open each VI in the LV 2011 it runs ok.
    Any ideas?
    How can i set TS dev system to use the 64-bit version? Or is there a separate 64-bit deployemrnt of TS?
    Labview 2010, TestStand 2010

  • I want to use Get Panel Image in Labview 5.0.1 and need details on how to save the BMP data generated

    I am trying to generate an application that saves a copy of its front panel on completion. This is easy to do using an invoke node with Print VI to HTML but this does not work in an .exe format. I have seen elsewhere that you have to use the Get Panel Image method, but no details are supplied in LV 5.0.1 documentation of how to use the "image" data (1-D Unsigned Byte array) that is generated. I want to save this in a format that can then be read as a bitmap in any standard graphics package. Any assistance?

    Hi,
    If you'd upgrade to LV5.1 or 6 you could use the 'standard' vi's for this.
    You need a VI called "Write BMP File.vi". It's not shipped with LV5.0.1.
    This vi only uses 3 subVI's, so perhaps someone at NI can convert it and
    send it to you (sorry, I won't, it's copywrited).
    If you cannot get this VI anywhere, you'll need to figure out the BMP file
    format yourself. It's not too complicated, but still could take some days.
    Perhaps someone figured it out before LV5.1 was released.
    Regards,
    Wiebe.
    "RDK" wrote in message
    news:[email protected]..
    > I want to use Get Panel Image in Labview 5.0.1 and need details on how
    > to save the BMP data generated
    >
    > I am trying to generate an application that saves a copy
    of its front
    > panel on completion. This is easy to do using an invoke node with
    > Print VI to HTML but this does not work in an .exe format. I have seen
    > elsewhere that you have to use the Get Panel Image method, but no
    > details are supplied in LV 5.0.1 documentation of how to use the
    > "image" data (1-D Unsigned Byte array) that is generated. I want to
    > save this in a format that can then be read as a bitmap in any
    > standard graphics package. Any assistance?

  • Labview 8 Application Builder and VISA Runtime Engine

    Hi,
    I am using Labview 8 Application Builder and everything is working fine except for 1 small thing. My program makes use of some of the VISA functions and when I build the application I'm having to include the VISA Runtime Engine Installer as a separate item on the CD to make the .exe file work.
    What I would ideally like is for people who want to use the application to run the installer and that all necessary components (including the VISA Runtime Engine) are also installed in the one installation. At the moment they're having to do 2 installations and I would like to streamline this.
    Is this possible?
    Ken

    Hi,
      this How-To explains about making those selections and including the different drivers :
    http://zone.ni.com/reference/en-XX/help/371361A-01/lvhowto/add_installers_to_build/
    and this tutorial covers the screenshots a little more step by step.
    http://zone.ni.com/devzone/conceptd.nsf/webmain/5ADBC06AC32E508A8625706E0062EBD1
    Hope that helps
    Sacha Emery
    National Instruments (UK)
    // it takes almost no time to rate an answer

  • How large can a LabVIEW Queue in elements and bytes?

    How large can a LabVIEW Queue in elements and bytes?

    rocke72 wrote:
    How large can a LabVIEW Queue in elements and bytes?
    In
    elements it is likely something like 2^31. In bytes it is most probably
    around the same number or better, depending how exactly the different
    queue elements are stored. I think they are stored as independent
    pointers so it would be theoretically better than those 2^31. In
    practice however starting to allocated so much memory in LabVIEW will
    always cause problems on nowadays computers. Without 64 bit CPU and OS
    going above 2 GB of memory per application is very hard and as far as I
    know not supported by LabVIEW.
    Also allocating many chunks of memory (like a few million queue
    elements holding strings) will definitely slow down your system tremendously eventhough
    it will work but the OS memory manager will be really stress tested
    then.
    Rolf Kalbermatter
    Message Edited by rolfk on 06-09-2006 12:24 AM
    Rolf Kalbermatter
    CIT Engineering Netherlands
    a division of Test & Measurement Solutions

  • Eye Diagram using Labview version 8.6 and Digitizer 5922

    I am trying to implement an eye diagram for some analysis. I have Labview version 8.6 and NI PXI Digitizer 5922. Is it possible to do some eye diagram analysis using this hardware/software. Do I need a specific toolkit for this?

    To answer your questions in order:
    Yes, you can and No you don't.
    Remember that an eye diagram is basically just an XY plot of one input channel against another. For example, amplitude vs phase.
    Mike...
    Certified Professional Instructor
    Certified LabVIEW Architect
    LabVIEW Champion
    "... after all, He's not a tame lion..."
    Be thinking ahead and mark your dance card for NI Week 2015 now: TS 6139 - Object Oriented First Steps

  • Mathematic​a Link for LabVIEW Integrates Design Models and Measuremen​ts

    Mathematica Link for LabVIEW, a new technical computing
    application created by BetterVIEW Consulting, is now available.
    This software bridges National Instruments LabVIEW graphical
    development environment and Wolfram Research Mathematica
    technical computing software to help engineers and scientists
    more easily acquire and analyze measurements throughout all
    phases of the design process.
    Mathematica Link for LabVIEW combines the data acquisition,
    analysis, and graphical user interface capabilities of LabVIEW
    and the modeling and analysis algorithms of Mathematica to speed
    development of custom applications in hundreds of industries,
    from quantum mechanics to automotive. With this software,
    engineers can control a LabVIEW application (Virtual Instrument
    or VI) from Mathematica or access Mathematica from within a
    LabVIEW VI.
    Mathematica Link for LabVIEW features built-in, high-level
    functions for creating a communication path between LabVIEW and
    Mathematica to achieve seamless integration throughout the
    design flow, from the first mathematical model to validating the
    final design. For example, engineers and scientists can easily
    send numeric computations to the Mathematica kernel and return
    results to LabVIEW, visualize LabVIEW data using native
    Mathematica data visualization functions, and create simple, yet
    flexible simulation and control VIs that take advantage of the
    symbolic programming capabilities of Mathematica.
    Integrating two leading technical computing systems like
    Mathematica and LabVIEW brings technical professionals one step
    closer to having the ideal, start-to-finish work environment.
    With Mathematica Link for LabVIEW, entirely new, hybrid
    workflows become possible, thus enabling users to quickly and
    easily find solutions to complex problems--solutions that would
    be difficult, if not impossible, to realize using only one of
    the component packages.
    Mathematica Link for LabVIEW requires Mathematica 4.1 or later
    and LabVIEW 6i or later, and is available for Windows and
    Macintosh platforms. For more information about Mathematica Link
    for LabVIEW, visit:
    http://www.wolfram.com/products/applications/labvi​ew

    Socalledvita-
    What functionality are you looking from the Amplitude and Level Measurement Express VI?  There is not one-one replacement you can use.  However, there are definitely ways of calculating the DC and RMS values of a signal as well as finding the max and min points, etc.  Please be more specific in what you are trying to accomplish.
    Regards,
    Mike S
    NI AE

Maybe you are looking for

  • ABAB RUNTIME ERROR IN BIT- TRAINIG

    Hi all          I try to apply for the support pack of Basis component in BI -Training system , During Apply it will stoped and give the Run time error . Then after am not able to execute any other tcodes inside the system . Below mentioned Error , K

  • What is iTunes unknown error -45054 and how do I get rid of it?

    I cannot download updates to apps because of this error.

  • Trying to use same device in subvi

    I have a main vi that takes a measurement using the daqmx read vi.  I would then like to take another reading from a different channel in a sub vi.  I am getting the 50103 resource is reserved error.  My initial thoughts were that the stop task was s

  • External swf not clickable

    I'm using addchild to load an external swf into a swfContainer, but i realize that all the buttons and text links doesnt seem to be working anymore.... swfContainer.addChild(loader); I try not loading it into a movieclip instead i load in on stage ad

  • Why won't my iSight webcam work with iChat?

    I have an iSight webcam and a Power Mac G5. I can't get the iSight to work with iChat, but it works with Yahoo Messenger, although the image is choppy. I have cable internet access, and I am connected to a wireless router (though not wirelessly, but