Labview and TI DSPs ?

Hi,
I bought (ouch.. that was years ago) a TMS320C5X ISA EVM. After having used
C5X assembly to write DSP applications, I was wondering if this card could
be used with NI LabVIEW (which I don't actually use though I would *really*
be interested in swaping asm programming for G language ;o), mainly for RT
(ANC) applications ? Are drivers (inc. generic TI DSP drivers) available or
do I have to link LabVIEW with Code Composer (via the TI DSP Toolset) ?
Thanks a lot for your answers...
Regards,
Nick.

Hi Nick,
With the DSP Test Integration Toolkit you will be able to interface with Code Composer Studio (meaning that all your programming will still be done in Code Composer Studio). What this toolkit does provide you is an easy tool to test your DSP code. You can automate the file management in Code Composer Studio and read to and write from RTDX Channels.
I hope this cleared your understanding. Please let me know if it didn't.

Similar Messages

  • Labview 6.0, rtdx com api, and ti dsp

    According to the ti code composer studio tutorial (ti document spru301c.pdf
    http://focus.ti.com/lit/ug/spru301c/spru301c.pdf, pg 1-13), one should be able
    to communicate between labview and a dsp using the rtdx com api. I know that NI offers
    the DSP Integration Kit, but it looks like some rtdx functionality can be accomplished
    without this by using this com api. I haven't been able to find any examples of doing this. TI
    did not seem to have any examples, I'm wondering if National Instruments has an example
    or can suggest a method of doing this.
    Thanks.
    regards,
    llp2

    Ame G.,
    Thanks for your suggestions. i tried the TI DSP Test example(Audio_dsk6711_bios_ccs2.20), but the example could not work correctly.
    The RTDX just transmitted the data from TI DSP C6711 DSK to Labview slowly. The data-rate is about 1 Hz. And i use the RTDX example in
    TI DSP tutorial to try again. I got the same result. I just received the data( 8 integrals) from TI DSP C6711 in 1 Hz. I want to use Labview
    to receive the data( 8 integral) in about 30 or 20 Hz. what is my problem? RTDX mode(continuous mode or non-continuous mode)? interrupt? or others?
    i provide my problem. Thank you very much and I am looking forward to hearing from you.
    regards,
    Danial 
    2005/10/20
    Attachments:
    less_1.zip ‏79 KB
    s1l1.c ‏3 KB
    RTDX.vi ‏97 KB

  • How to switch central frequency fastly with labview and USRP?

    Hello everybody,
             Recently I was making a frequency hopping receiver with labview and USRP N210 and SBX. The transmitter works on the 2.4G band, and it changes from one central frequency to another every 65536us. As I think, if I start receiving at the correct time, and  change the central frequency of the USRP every 65536us, I should be able to receive the data.
        But I find that, when usrp switch form one central frequency to a different one, it spends nearly 48ms and when it switch from one frequency to the same one, it only spends 3ms(I test the performance by modifying the niUSRP EX Rx Continuous Async Reconfig on the Fly.vi example). So if I want to follow the hopping sequency, I may can only receive 17ms data because I have to save time for switching central frequency. 
         My problem is that: why usrp changes central frequency so slow,  how can I make the process faster? If I want to follow the hopping sequency, can anybody give me some good advice? Thanks

    There is an attribute "LO Frequency" that you can use to lock down the frequency of the local oscillator.  Then, when you set the "Carrier Frequency" attribute, the difference between the carrier frequency value and the LO frequency value will be realized by frequency shift in the DSP in the FPGA.  Changing this "digital" frequency shift value is much faster, since you don't have to re-tune the local oscillator.
    Let's say you're sweeping from 2.44 GHz to 2.46 GHz.  Set the "LO Frequency" to 2.45 GHz then Commit.  Then, in your loop, program the "Carrier Frequency" attribute from 2.44 GHz to 2.46 GHz.  In the loop, it will only program the digital frequency shift, which will happen much faster.
    Another tip is to use the property node to program only the attributes you want to change, rather than the niUSRP Configure Signal.vi (which sets many at once).
    For more details, see the "Configuration:Advanced:LO Frequency Property" in the NI-USRP Help, and look at this example:
    <LabVIEW dir>\examples\instr\niUSRP\niUSRP EX Fine-Tune LO Frequency.vi

  • I am receiving the data through the rs232 in labview and i have to store the data in to the word file only if there is a change in the data and we have to scan the data continuasly how can i do that.

    i am receiving the data through the rs232 in labview and i have to store the data in to the word or text file only if there is a change in the data. I have to scan the data continuasly. how can i do that. I was able to store the data into the text or word file but could not be able to do it.  I am gettting the data from rs232 interms of 0 or 1.  and i have to print it only if thereis a change in data from 0 to 1. if i use if-loop , each as much time there is 0 or 1 is there that much time the data gets printed. i dont know how to do this program please help me if anybody knows the answer

    I have attatched the vi.  Here in this it receives the data from rs232 as string and converted into binery. and indicated in led also normally if the data 1 comes then the led's will be off.  suppose if 0 comes the corresponding data status is wrtten into the text file.  But here the problem is the same data will be printed many number of times.  so i have to make it like if there is a transition from 1 to o then only print it once.  how to do it.  I am doing this from few weeks please reply if you know the answer immediatly
    thanking you 
    Attachments:
    MOTORTESTJIG.vi ‏729 KB

  • Problem with Labview and an ARM Cortex

    Good morning,
    I am currently trying to use Labview with a board from ST Microelectronics (MCBSTM32) with an ARM Cortex Processor.
    I use the SDK and have followed the tutorials.
    But, when I try to launch the program (the simple loop as write in the tutorial n°2 : http://zone.ni.com/devzone/cda/tut/p/id/7029 ), Keil gives me an error via Labview:
    "Argument 'DARMSTM' not permitted for option 'device'."
    It seems that Keil does not allow an ARM Cortex from ST as the device.
    Moreover, after having this problem, I am unable to use a Keil project ,even a project which worked before, without Labview. I need to restart the computer.
    I also try to launch the Keil project generated by Labview, without using Labview, and it works. But as soon as I use Labview, I have the error.
    Did anyone already have this error or know how to solve it
    Thank you for your answer and sorry for my bad english.
    Regards,
    Raphaël VAISSIERE

    Hi Raphi,
    So let me make sure I understand,
    The project created in LabVIEW errors out with the message "Argument 'DARMSTM' not permitted for option 'device" 
    If you open the same project in Keil uVision, it runs fine
    Here are my questions:
    1. So how does the code run when run through Keil? does it deploy and run fine?
    2. Did you follow the porting procedure completely?
    Your target STM32F103RB  is techincally supported by Keil but you need to port the RTX kernel to it. This paragraph explains it:
    To determine if your target already supports the RTX Real-Time Kernel, browse to the \Keil\ARM\Startup directory, then browse to the folder that corresponds to the manufacturer of your ARM microcontroller. If there is an RTX_Conf*.c file for your target, then the RTX Real-Time Kernel has already been ported for your ARM device. If no such file exists, skip to chapter 4 for more information on the RTX Real-Time Kernel and a guide for porting RTX to your ARM microcontroller. 
    You also need to port the Real-Time agent to it.
    I just want to make sure that you have followed the guidelines. If you have and are still having problems, we will continue to explore this.
    Thanks,
    National Instruments
    LabVIEW Embedded Product Support Engineer

  • I am trying to use Labview and RP1210 compliant hardware to connect to a truck J1939 bus and receive messages.

    I am trying to use Labview and RP1210 compliant hardware to connect to a truck J1939 bus and receive messages. 
    Specifically I am attempting  to read data frames using the RP1210_READMESSAGE .   am able to configure the hardware and send a message to the J1939 bus. .    I think I have not configured something correctly.  I can use the RP1210_SENDMESSAGE and see the message I have sent on the bus using CANalyzer   When I use the RP1210_READMESSAGE   I get the timestamp from a message and the return from the function sends back the correct number of bytes (the number matches the number of bytes I sent out plus four bytes from the timestamp).  What I am having trouble with is actually receiving the data. I have had the same type of behavior from two different hardware (Vector CANcase XL and Nexiq USB Link), so I don't think the issue is vendor specific.
    Has anyone been able to make the RP1210_RECIEVEMESSAGE function work correctly?
    Thanks for any help

    Thanks
    I have already tried that.  The links are the NI RP1210 wraper. The problem I am having is using labview to interface with the RP1210 layer.  The RecieveMessage char*fpchAPIMessage this is the output which is a pointer to a cahracter array.  In this variable I can receive the timestamp of the message but not the message.  The retun showns the correct amount of bytes are aviaable (18 for a 8 byte message) but I can only get the 4 byte timestamp  I think I have to dereference this pointer to view the data.  I am not sure how to fix this. 

  • Tcp data b/w labview and c++

    Hi
     i am trying to establish TCP connection b/w LABVIEW and C++ program. Server is established in C++ while client is implemented in labview.Although connection is successfully establish b/w server and client, both are unable to correctly understand data send/receive among them. Forexample if i want to send an int type send_array from server, i use standard WINSOCK function "send" like that:
    send(AcceptSocket,(char*)send_array,129*4,0);
    but when client in labview receive this array, it shows unexpected values.As a client, I used "simple data client.vi"  with one modification i.e. as sent data size(129*4 bytes) was fixed, only one TCP read was used . 
    Same problem exists if i send data from client to server.
    Kindly help me
    Best Regards
    Solved!
    Go to Solution.

    It's probably an big/little edian problem. If you are using the Flatten/Unflatten from String functions, you can specify which to use.

  • Questions on Saving and mining data with Labview and DIAdem

    Hi,
    I am sampling two signals at 200k sampling rate. I am trying to save the data on harddisk and analyzing the data using DIAdem.
    If I save the data using mesurement file format .tdm and .lvm, the file size will be about 4 Gigabytes for only 10 mins' acquisition. It is very slow to process it.
    I used the software Clampex and pCLAMP(Axon Instruments) before. At the same 200k sampling rate and also acquiring two signals, these programs save the data as .atf format and the size is only 400 MB for 10 mins' acquisition.
    I wonder if there is also a good way to handle this situation using Labview and DIAdem, and how to do it?

    Hi, Jonathan:
    I tried the TDM binary file format. The file size is 800 MB per min acquisition and it will take a long time to just open these files. For my application, I have to take data for several hours. I am looking for a way to reduce the size of files.
    Is there any other type of file format that can reduce the size significantly and can be handled easily?

  • Please Help!!! VIPM won't connect to labview and I need to figure out why

    I am trying to download the Arduino package from the VIPM and they won't seem to connect. I have included a picture of the error and two more of my settings. I am using the student evaluation of the Labview software, but I'm not sure that it is what's causing the problems. Any help would be appreciated, as I would like to set up a nice GUI using labview and an arduino as my DAQ device. I need this for a project for my professors so I am rather desperate to get it up and running.
    Attachments:
    Labview Error.jpg ‏352 KB
    Labview Error II.jpg ‏141 KB
    Labview Error III.jpg ‏108 KB

    No, i tried using you settings and stil not working. i went through every combination of the four items I have on the machine and export list. Still nothing, do you think that its that i am using the student version?
    Attachments:
    New Labview settings, not working.jpg ‏123 KB
    New Labview settings, not working.png ‏267 KB

  • Can I use a camera for application in Labview and VBAI at the same time ?

    Dear all,
    I'm trying to save an AVI file with Labview and make an image process with VBAI at the same time, in one machine.
    The error : "Camera already in use" displayed.
    My Camera is a GIGE and I work with Imaqdx. I've test the multicast mode but it only operate with several machines.
    How can I do this ?
    Thank's to help me,
    Yoann B

    I'm not necessarily saying that.  It's been a while since I've used VBAI, so I don't remember all of the capabilities, but if VBAI can do the inspection and recording at the same time, you should be fine.
    The trick is that only one program can access the camera at the same time.  That application reserves the camera, thus making it unavailable to others.
    Chris
    Certified LabVIEW Architect
    Certified TestStand Architect

  • Tdms-files: Is it possible to prevent LabVIEW and DIAdem from creating .tdms_index files?

    Hello,
    is it possible to prevent LabVIEW and DIAdem from creating .tdms_index files when opening/creating/editing a .tdms file?
    I think I have no benefit from the .tdms_index files because our applications create a lot of little .tdms-files (repeat measurements). With the additional .tdms_index files we have unnecessary memory consumption and it takes longer for Windows to open the containing folder. Also it´s confusing when searching for a certain file.
    Best Regards
    Daniel
    Solved!
    Go to Solution.

    Hi Baui,
    I'm afraid there's a direct way to disable creating .tdms_index file now from LabVIEW and DIAdem. You can use TDMS Advacend API in LabVIEW, which doesn't create any index file, or you can just make some simple programming and delete the .tdms_index file after closing the TDMS file. 
    This is a request for a long time, we'll consider to add this feature for TDMS in future releases.
    Yongqing Ye
    NI R&D

  • Problems with Labview and Windows XP with error -214721786​5

    Hello. Im a new beginner in Labview and am having a problem with a error message
    I dont understand. Im running Labview on a pc-logger with windows XP.
    The program is written by a guy on my university, and logs voltage from
    a solar cell panel. He doesnt have the time to check in to it, so Im supposed to
    find out what's wrong.
    The error message is as follows :
    Error -2147217865 occured at Cmd Execute.vi->DB Tools Insert
    Data.vi->Solcelle.vi
    Possible reasons :
    Exception occured in Microsoft OLE DB Provider for ODBC Drivers,
    [MySQL][ODBC 3.51 Driver][mysqld-3.23.38-log]Can't open file :
    'solcelledata.MYD'. (errno: 145). in Rec Create - Command.vi-
    >Cmd Execute.vi->DB Tools Insert Data.vi->Solcelle.vi
    I really hope someone can tell me what's wrong and specifically what to do
    with it, or possibly guide me to someone who could suggest a solution to this
    problem.
    Reguards
    chrisdee

    Hi Chrisdee,
          (.. not to discount what TimoS has suggested, however...) 
    Embedded in the error information you provided:
    ... Can't open file : 'solcelledata.MYD'.
    Sounds like the program needs this file?  Has the program been moved to a new location?
    just a wild guess!
    Cheers
    Message Edited by Dynamik on 10-06-2005 04:53 AM
    When they give imbeciles handicap-parking, I won't have so far to walk!

  • Hi! can any1 pls let me know how to create a look up table in labview and use that in the vi.

    hello !
    i have no idea how to build a lookup table(as v use in microcontroller) in labview and use that in vi.pls help me
    txs
    nitin
    Solved!
    Go to Solution.

    If the lookup table is always going to remain the same (e.g. a character generator or something similar) you can place the values in a 2D array constant on your diagram, with the input value as one column, the equivalent as the other. When you need to perform the lookup you use an index array to return all the values in the "input column", search it using "search 1D array" and use the resulting index number to index the other column's data. If the values may change, then it would probably be best to load an array control with your equivalent values from a file.
    P.M.
    Putnam
    Certified LabVIEW Developer
    Senior Test Engineer
    Currently using LV 6.1-LabVIEW 2012, RT8.5
    LabVIEW Champion

  • Are there any 3rd party LabVIEW and/or TestStand code analysis tools

    I am looking for any 3rd party tools to analyze LabVIEW and/or TestStand code. Looking for tools that will check for memory leaks, timing, and any other code performance problems. I know LabVIEW has the profiler tool, but I am looking for 3rd party tools.

    Hi wbolton,
    I am personally not aware of any 3rd party tools that do the same thing as the Profiler, but I can say that if you'd like to analyze a .VI or .seq file, the most straightforward way to do so would be with the appropriate National Instruments software.
    Regards,
    Dan Richards
    Dan Richards
    Certified LabVIEW Developer

  • Interfacin​g between labview and microcontr​oller using rs 232

    hi my task is to generate a wavefrom of certain frequency and certain duty cycle in labview and feed it to pic18f452 microcontroller and the o/p of microcontroller when connected to CRO should display the same frequency and duty cycle which was fed from labview.
    for that when i am sending decimal data from labview microcontrolle is receiving ascii data for eg 62(decimal) gives 36 and 32 respectively in microcontroller registers and when i use type cast function and convert the decimal data into ascii data and then sending it to uc it gives hex o/p so i am confused that how to solve this problem
    pls guide me
    thanks
    satish

    So your display from the uc is giving you a hex display instead of the ASCII representation? Are you using VISA to send a string to the uc? Is that string being represented as hex instead of ASCII by the microcontroller? I'm confused about what exactly your setup and problem is...

Maybe you are looking for