Xbee data extracting and graphing

Hi,
I have been trying for a few days to get this to work with no results. I have searched through google and some of the forums for help with nothing that really helped.
I am using two xbee devices to transfer data to labview to graph.Eventually I would like to use the ADC on the Xbee to take an analog signal and convert it to a digital signal for labview to read and so I can do my analysis on it. I started with getting a basic Visa program to work and send number from my arduino uno over and over again and placing those values into an array This works great with no problems. The problem comes when I try and send a signal through the Xbee. 
I tried to send an analog signal out to the adc and I was getting some values in labview in Hex but I was unsure what they really meant. My second attemp was just to try and create a pulse wave using my arduino and send that to labview and have labview try and graph the incoming signal. This however did not work and I am not getting any data. I connected both Xbees up on a bread board without labview to see if I can get the a signal out of the recieving xbee and it worked. So the problem should not be with my xbee's but with my code. 
The code is where I am not sure about. I was able to find examples of how to use the visa and was given a basic example of it. I have then tried to manipulate the program to see if ti would work for me. I am trying to take the signal from the VISA Read funciton and convert it to a 1d array to display on a graph. I have also tried to seperate the dat into an array char by char from the string. However, none of this worked. I am getting a zero value out for my graph and no values into my array. I am also getting no value into my READ String area. 
The values I have sometimes gotten look like:
00\00\00\00\00\94\00\00\ and so on with different value. Some time the byte read is very large up to 475 when I am placing the pulse wave in or so and sometimes smaller at 19 when I tried to connect the ADC.
I have attached some of my VI's to see if anyone can come up with a better way or point me in a better direction.
The png is a pic of the results I got from the pulse input. So it looks like I am not getting any data at all. 
Thanks for the help
Juan
Attachments:
Basic VISA_Test_1.vi ‏26 KB
Result_OfpulseWave.png ‏129 KB

Hello jrod03,
I would suggest using the NI LabVIEW Interface for Arduino Toolkit found here (http://sine.ni.com/nips/cds/view/p/lang/en/nid/209​835) and posting your Arduino related questions in our LabVIEW Interface for Arduino community (https://decibel.ni.com/content/groups/labview-inte​rface-for-arduino).
This document will also probably be helpful to you (http://digital.ni.com/public.nsf/allkb/8C077471896​06D148625789C005C2DD6).
Jonathan L.
Applications Engineer
National Instruments

Similar Messages

  • Data Visualization and Graphs for JSF

    Are the Data Visualization and Graphs for JSF included in this release. I saw the demo here http://www.oracle.com/technology/products/jdev/11/index.html and want to try it out. But I could not seem to access it in the tool.
    Regards
    Orlando Kelly
    Cayman Islands Government

    <p>
    Hi,
    </p>
    <p>
    It seems you have no suitable libraries in your ViewController project. See on this image whet libraries you need.
    </p>
    <p>
    Kuba 
    </p>

  • Data Extraction and ODS/Cube loading: New date key field added

    Good morning.
    Your expert advise is required with the following:
    1. A data extract was done previously from a source with a full upload to the ODS and cube. An event is triggered from the source when data is available and then the process chain will first clear all the data in the ODS and cube and then reload, activate etc.
    2. In the ODS, the 'forecast period' field was now moved from data fields to 'Key field' as the user would like to report per period in future. The source will in future only provide the data for a specific period and not all the data as before.
    3) Data must be appended in future.
    4) the current InfoPackage in the ODS is a full upload.
    5) The 'old' data in the ODS and cube must not be deleted as the source cannot provide it again. They will report on the data per forecast period key in future.
    I am not sure what to do in BW as far as the InfoPackages are concerned, loading the data and updating the cube.
    My questions are:
    Q1) How will I ensure that BW will append the data for each forecast period to the ODS and cube in future? What do I check in the InfoPackages?
    Q2) I have now removed the process chain event that used to delete the data in the ODS and cube before reloading it again. Was that the right thing to do?
    Your assistance will be highly appreciated. Thanks
    Cornelius Faurie

    Hi Cornelius,
    Q1) How will I ensure that BW will append the data for each forecast period to the ODS and cube in future? What do I check in the InfoPackages?
    -->> Try to load data into ODS in Overwrite mode full update asbefore(adds new records and changes previous records with latest). Pust delta from this ODS to CUBE.
    If existing ODS loading in addition, introduce one more ODS with same granularity of source and load in Overwrite mode if possible delta or Full and push delta only subsequently.
    Q2) I have now removed the process chain event that used to delete the data in the ODS and cube before reloading it again. Was that the right thing to do?
    --> Yes, It is correct. Otherwise you will loose historic data.
    Hope it Helps
    Srini

  • Help In keithley 2400 VI!!(Problem with the data logging and graph plotting)

    Hi,need help badly=(.
    My program works fine when i run it,and tested it out with a simple diode.The expected start current steps up nicely to the stop current.The only problem is when it ends,i cannot get the data log and the graph,though i already have write code for it.Can someone help me see what's wrong with the code?I've attached the necessary file below,and i'm working with Labview 7.1.
    Thanks in advance!!!
    Attachments:
    24xx Swp-I Meas-V gpib.llb ‏687 KB

    Good morning,
    Without the instrument it might be hard for others to help
    troubleshoot the problem.  Was there a
    specific LabVIEW programming question you had, are you having problems with the
    instrument communication, are there errors? 
    I’d like to help, but could you provide some more specific information
    on what problems you are encountering, and maybe accompany that with a simple
    example which demonstrates the behavior? 
    In general we don’t we will be unable to open specific code and debug,
    but I’d be happy to help with specific questions. 
    I did notice, though, that in your logging VI you have at
    least one section of code which appears to not do anything.  It could be that a small section of code, or
    a wire was removed and the data is not being updated correctly (see pic below).  Is your file being opened properly?  Is the data being passed to the file
    properly?  What are some of the things
    you have examined so far?
    Sorry I could not provide the ‘fix’, but I’m confident that
    we can help.  Thanks for posting, and
    have a great day-
    Message Edited by Travis M. on 07-11-2006 08:51 AM
    Travis M
    LabVIEW R&D
    National Instruments
    Attachments:
    untitled.JPG ‏88 KB

  • Need data extract and load program

    hi
    where i can find all the standard data extract programs and standard load programs. Pls thisis urgent.
    thankyou

    Thankyou
    can i get a list of the extract programs and dataload programs? SAP provides programs for extracting data for material master, customer, vendor, purchase order , sales order ? It also provides std programs for data load rite? where can i find the list. i need to prepare a document with the list.
    Heng

  • QUESTION:  Essbase data extraction and Installing ODI Agent??

    For extracting data from Essbase cubes, ODI has "LKM Hyperion Essbase DATA to SQL".
    We can use (1). ReportScript, or (2). MDX-query, or (3). CalcScript
    For data-extraction using CalcScript, ODI Agent must be running on the same server as the Essbase server.
    Does anyone know if there is a need for ODI Agent on the Essbase machine if we use MDX-query method for data-extraction?
    We would like to avoid installing ODI Agent for Essbase data-extraction.
    .

    Thanks John.
    One related question. To move data from one Essbase cube to another Essbase cube using ODI Interface, Can we do it efficiently through MDX-query?
    We want to avoid Replicated-partitioning OR CalcScripts, if possible.
    BTW... Your ODI/Hyperion blog is a bible for us.

  • Byte data extraction and interpretation from a WAVE sound file

    Hello,
    I'm trying to extract the signal data from a WAVE-file for further processing. Suppose the WAVE-file is in 16 bit/stereo format - then I get (8 bit) low-byte and (8 bit) high-byte for the left channel first, and then the low- and high-order bytes for the right channel (I hope I've got it right this far).
    Now since I'm interested in doing some frequency analysis on the signal, I need to have the amplitudes for every sample, but since each byte is signed I can't just compose the complete amplitude value from the two bytes for each channel. A simple way to solve this would be (I think) to check if a byte is negative, and if it is, store the absolute value of the byte + 127 in an integer. For example, the byte 11111111 would be interpreted as -128, but taking the absolute value and adding 127 produces 255 which would be the maximum "unsigned" value for an 8 bit binary number.
    This seems like a quite "ugly" solution to me, and I don't really like ugly things. I haven't worked that much with binary math in Java, and my general knowledge on the subject is not very thorough in the first place. I would be very grateful for any suggestions.

    Ok, I went to buy some food and came up with another solution:
    Suppose I have a sample from one of the channels in my byte array "signal" and I want to compose a (signed) total amplitude of the two bytes in the array. Code:
    byte[] signal = ... ; //Get the data
    int amp = 0x0000;
    amp = ((signal[1] & 0xFF) << 8) | (signal[0] & 0xFF);
    Here I assume that the data is aquired in little-endian format. If you have a better solution, please let me know.

  • Data handling and graphs

    I originally wrote the graph program to handle the data in a text file that was organized like this...
    vertex a
    vertex b
    vertex c
    vertex d
    edge a c
    edge a d
    edge d b
    and now I have to change the main to accept a datafile containing...
    a b
    b c
    c e
    d g
    g c
    Now, here is a copy of the main program as it currently stands...
    import java.io.*;
    import java.util.*;
    public class TopSort
         static Graph theGraph = new Graph();
         static Hashtable hashList = new Hashtable();  //just to store array index
         public static void main (String args[])
                 MyInfoachaffin myInfo = new MyInfoachaffin();
              myInfo.info();     
                 File sourceFile = new File(args[0]); // access the file
                 if (sourceFile.exists() == false)                                          
                   System.err.println("Oops: " + sourceFile + ": No such file");
                   System.exit(1);
                 String newVertex, startEdge, endEdge;
                 int arrayPosition = -1;
                 try //open the file
                           FileReader fr = new FileReader(sourceFile);
                           BufferedReader br = new BufferedReader(fr);
                        String input;
                        while ((input = br.readLine()) != null)
                             StringTokenizer toke = new StringTokenizer(input);
                                while (toke.hasMoreTokens())
                                  if (hashList.containsValue(toke))
                                               return;
                                     else
                                          newVertex = toke.nextToken(); //get vertex
                                          theGraph.addVertex(newVertex); //add into graph
                                          arrayPosition++; //increment counter
                                          hashList.put(newVertex, new Integer(arrayPosition));
                                          //add position with vertex as key
                                     /*else if (toke1.equals("edge"))
                                          startEdge = toke.nextToken(); //get edge
                                          endEdge = toke.nextToken();  //get vertex
                                          Integer temp = ((Integer)hashList.get(startEdge));
                                          int start = temp.intValue(); //find position with key
                                       Integer temp2 = ((Integer)hashList.get(endEdge));
                                          int end = temp2.intValue();  //find position with key
                                          theGraph.addEdge(start, end); //add edge
                                }//close inner while
                       }//close outer while
                       System.out.println("The hashtable contents: " + hashList.entrySet());
                        br.close();
                  }//close try
                      catch (IOException e)
                                System.out.println("Whoops, there's a mistake somewhere.");
                          theGraph.graphSort();
       }//end main
    }//end class I have managed to seperate the vertexes from the edges and stored them in a hashtable so as to be able to remember their location in the array itself. My question is is there a way to go through a file a second time and pull the vertexes from that file or, conversely, should I store the data in the file into, oh I dunno, a linked list or something that I can pull the data from? The graph part of the program works so I didn't add that file and the setting the vertexes works fine too...I'm just stuck on how to handle the edges, exactly, so does anyone have any advice? Right now, the adding edges is commented out because that was how I handled it with the original data...

    Whoa, you're freakin' me out.
    All you gotta do is read in the data in a new format. You can either translate the new format to the old format, or you can write a method that creates objects from the new format explicitly. That's all there is to it.

  • Data Extract and Data Load Through EPMA

    Hi All,
    Hope you all are doing well,
    Can we Extract Data from classic application and Load Data into another classic application through EPMA, Is it written in any document of oracle.
    If yes then please help me its urgent.
    Thanks,
    Avneet
    Edited by: Avneet on Mar 16, 2011 1:31 AM
    Edited by: Avneet on Mar 16, 2011 2:29 AM

    Hi John,
    I have tried a lot using ODI, i have raised some forum also but i am not able to do it through ODI, in ODI i am using it through report script and getting some error
    Re: How to extract data From Hyperion Essbase to Flat Files
    Thanks,
    Avneet

  • XML data extraction and new row display

    I have a table named, SAPDATABROWSER, which has multiple columns and rows. A column named SAPTEXT (DataType: CLOB) consist of xml data. Each XML data is different in terms of number of nodes, but has the same following structure:
    <SAP_BAPI_PROP>
    <SelectedProperty>
    <Structure>
    <Field></Field>
    <Length></Length>
    <FieldValue></FieldValue>
    <Type></Type>
    <Priority></Priority>
    </Structure>
    </SelectedProperty>
    </SAP_BAPI_PROP>
    So now what I would like to do is, to extract the values: Field, FieldValue and Length from every row (total 72) and display the result as following,
    Field FieldValue Length
    XXXX YYYY ZZZZ
    PPPP QQQQ RRRR
    AAAA BBBB CCCC
    MMMM NNNN OOOO
    Again, each row (total 72) has multiple nodes inside XML data, i.e there are multiple Field,Length and FieldType Nodes and no two XML row/data are similar!
    I am using the following code:
    SELECT
    EXTRACT (xmltype (saptext), '/SAP_BAPI_PROP/SelectedProperty/Structure/Field/text()'),
    EXTRACT (xmltype (saptext), '/SAP_BAPI_PROP/SelectedProperty/Structure/Length/text()' ),
    EXTRACT (xmltype (saptext), '/SAP_BAPI_PROP/SelectedProperty/Structure/FieldValue/text()' )
    FROM sapdatabrowser ;
    The above code generate result as following:
    /SAP_BAPI_PROP/..../Field/text() /SAP_BAPI_PROP/..../Length/text() /SAP_BAPI_PROP/..../FieldValue/text()
    1. XXXXPPPPAAAA YYYYQQQQBBBB ZZZZRRRRCCCC
    2. MMMM NNNN OOOO
    Any suggestions?! Thank you very much. :)

    Thank you everyone for your input. But I am still stuck. Let me start with my SQL version:
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    PL/SQL Release 11.2.0.1.0 - Production
    CORE     11.2.0.1.0     Production
    TNS for 32-bit Windows: Version 11.2.0.1.0 - Production
    NLSRTL Version 11.2.0.1.0 - ProductionI believe my explaination was not clear so let me rephrase my question. Following is a screenshot of the Table SAPDATABROWSER
    http://img585.imageshack.us/img585/2724/55760509.jpgSo there are total 72 rows, and my interest is in Column 6, SAPTEXT (type:CLOB). Each of the row has following structure:
    <SelectedProperty>
    <Structure>
    <Field></Field>
    <Length></Length>
    <FieldValue></FieldValue>
    <Type></Type>
    <Priority></Priority>
    </Structure>
    </SelectedProperty>Each row has multiple nodes Field, Length, FieldValue.... etc. Now I would like to extract ALL: Field, Length and FieldValue nodes from ALL rows and display them in following format
    Field                  Length                   FieldValue
      1                       1                           1
      2                       2                           2
      3                       3                           3
      4                       4                           4Example Output:
    ROW 1:
    <SelectedProperty>
    <Structure>
    <Field>X</Field>
    <Length>Y</Length>
    <FieldValue>Z</FieldValue>
    <Type></Type>
    <Priority></Priority>
    </Structure>
    <Structure>
    <Field>P</Field>
    <Length>Q</Length>
    <FieldValue>R</FieldValue>
    <Type></Type>
    <Priority></Priority>
    </Structure>
    </SelectedProperty>
    ROW 2:
    <SelectedProperty>
    <Structure>
    <Field>A</Field>
    <Length>B</Length>
    <FieldValue>C</FieldValue>
    <Type></Type>
    <Priority></Priority>
    </Structure>
    <Structure>
    <Field>M</Field>
    <Length>N</Length>
    <FieldValue>O</FieldValue>
    <Type></Type>
    <Priority></Priority>
    </Structure>
    </SelectedProperty>Required Output:
    Field                  Length                   FieldValue
      X                       Y                           Z
      P                       Q                           R
      A                       B                           C
      M                       N                           O

  • Actual Cost, Resource and WBS data extract and upload

    Can anyone help with getting data out of Project system?  Does an extract/upload BAPI / IDoc exist?
    If not does anyone have any other advice on how to do this?
    Rgds,
    Mark

    just go to the tables and extract to excel or use reports to extract data to spreadsheets

  • Master data Extraction and Repeat delta

    hi Expert's I am extracting master data from R/3.These are Delta loads.Now when I started extracting it is pops up message stating earlier request is not correct so repeat must be done.could any one of analyse the problem and give the solution.I also dont know how to do repeat delta.Plz explain with steps.
    Thanks in Advance.
    vasu.

    Hi Vasu,
    Goto Manage, Select the Request and make it red(See below how to make it red), then click on the Monitor icon(just side to it). It will take you to Monitor screen.
    here select the particular load, on the right side, on the status tab, click on the load and make it red.
    Now goback to the infopackage, schedule, it will do repeat delta.If the Problem still persist, then delete the Requet in the Manage screen and schedule again. This should work
    Regards
    Vj

  • Generic data extraction and loading to cube

    Hello guys, I am trying to load R3 data thru Gernic extraction to infocube in bw3.5.
    I just created generic DS and activated it and also Replicated in BW side.
    I am trying to extract transaction data - from FI , DS has Ship Info Table fields in the extraction structure.
    Everything got activated and now i am in BW side, just assigned DS to infosource.
    But i am stuck in Transfer rules, all infoobjects on the left are empty and it wont let me activate.
    I am really confused, Should i be creating infoobjects first , because i havent created infocube yet.
    My DS has about 12 fields , I really need someone to tell me why its not activating Transfer rules, Comm structure is still empty.
    Should i be creating INfocube first with KF and Chars?or how am i supposed to map these DS fields in the Infocube (my data target)?
    I guess i am not clear at this point, even with LO extraction I could get create DS and replicate in BW , but once i am in BW i become clueless i should say.I did search other posts here but couldnt find anything that help me understand this.if someone could explain me in a simple terms i would appreciate it.
    I have created DS with extraction structure, i am in BW , so whatever field i selected in extract structure in R3 side f rom the table , is that going to be Chars and Key figures in my Infocube once i complete Loading ? or how would that work? Why would we need to chose specific Table when creating view , assuming that we already know the data that we need for reporting purposes, we would know those data should be on what table correct?
    please drop some lines to help me thanks

    hello again, i am writing this since i didnt get any response. I would really appreciate it if someone could give me little hint , i have been practicing on my own using bw system and need help from you Pros.
    My previous question was regarding Transfer rule . I am still not able to get through that steps.
    This is what i have done so far:
    -created DS generic for transaction data,  using View VBAK table (not sure if this was the right table, i just wanted to see if i would be successful loading data in the cube).
    - ACtivated DS, replicated in BW, Assigned INfosource.
    - I selected 15 Fields from DS (extraction structure ) from that VBAK table.
    - But when i am in Transfer rule ./structure screen, there are many fields appearing.
    - It will let me active the Transfer Rules, however i also created InfoCube , it asked me to chose at least one Time Chars, and KF. But i used Template using my infosource, there were no KFs . so i figured i need to change Transfer rule. I tried to create Key figure infoobject on the right side of the screen in Transfer rules, but it will not move to the right when i tried to transfer it.
    My question is, why there are more fields in this screen and some fields are not appearing the one i selected.
    Since i chose transaction Generic DS,is that why i have to have Key figures? did i chose worng table VBAK for practice purpose? i dont really see much Keyfigures when i looked the content of this table using SE16.
    Guys please suggest me , what route should i take to get through this confusion.
    My main objective here is to simply load R3 data using Genric DS(transacitonal)  into a customized infocube.
    I get error when creating Update rule, infosource and infocube wouldn not link, i think its becoz i didnt have any keyfigures available when creating Infocube ? how would i chose one?
    anyone please throw me some suggestions ..i would really appreciate it.thanks again for reading

  • Open data extraction orders -  Applying Support Packs

    Dear All,
    I have done the IDES 4.6C SR2 installation.
    While updating the support packs, i get the message saying
    CHECK_REQUIREMENTS phase.
    Open data extraction orders
    There are still open data extraction orders in the system
    process these before the start of the object import because changes to the ABAP Dictionary structures could lead to data extraction orders not being able to be read after the import and their processing terminating
    For more details about this problem, see Note 328181.
    Go to the Customizing cockpit for data extraction and start the processing of all open extraction orders.
    I have checked the Note.
    But this is something m facing for the first time.
    Any suggestion!!!
    Rgds,
    NK

    The exact message is :
    Phase CHECK_REQUIREMENTS: Explanation of the Errors
    Open Data Extraction Requests
    The system has found a number of open data extraction requests. These
    should be processed before starting the object import process, as
    changes to DDIC structures could prevent data extraction requests from
    being read after the import, thus causing them to terminate. You can
    find more information about this problem in SAP Note 328181.
    Call the Customizing Cockpit data extraction transaction and process all
    open extraction requests.

  • Difference between Data staging and Dimension Table ?

    Difference between Data staging  and Dimension Table ?

    Data Staging:
    Data extraction and transformation is done here.
    Meaning that, if we have source data in flat file, we extract it and load into staging tables, we take care of nulls, we change datetime format etc.. and after such cleansing/transformation at then end, load it to Dim/Fact tables
    Pros: Makes process simpler and easy and also we can keep track of data as we have data in staging
    Cons: Staging tables need space hence need memory space
    Dimension Table:
    tables which describes/stores the attribute about specific objects
    Below is star schema which has dimension storing information related to Product, Customer etc..
    -Vaibhav Chaudhari

Maybe you are looking for

  • I want to delete IP address in the DHCP manager(Windows NT) by VIs

    Some one please advise this question.

  • Closed laptop sleep mode

    Is it safe to just close the laptop and put it in sleep mode for moving it around o taking it to work?, some people say is unsafe because the hard drive, as the only moving mechanical part, can get damage because of this movement. I think it`s safe b

  • Compatibility of Essbase V11.1.2.2

    Hi All, We have a plan to upgrade the essbase from V11.1.1.3 to V11.1.2.2. could you please let me know the - compatability of Essbasae V11.1.2.2 with SQL Server 2008 R2? - compatability of Essbasae V11.1.2.2 on OEL 5.3 operating system. and also I h

  • Making a silent film

    I'm making a silent film in iMovie HD and I need to create a "title template" that has an antique border where I can type in the dialogue of the characters. In many of the old silent films, they would cut from the characters speaking to a black "titl

  • Offline adobe forms using webservices-how to print table data

    HI, i have created offline adobe form using webservices which call rfc to pull the data ...to get inspection lot details ... iam getting header data....but iam not able to get multiple line items of the table data ..here in scenario there are 4 line