Data logging using Blue Thooth

While we are waiting for the Mindstorm toolkit for LV to apear.
Could I write a program in Lego Mindstorm NXT software, that uses Blue Thooth to send data.
And then write a program in LabView collecting this data?

Hi Kjeld,
I haven't yet tried to break it, but here are the steps I used to get the Simple Bluetooth Server.vi/Simple Bluetooth Client.vi to work...
1) I inserted a BT/USB dongle in a USB port of my desktop computer running WinXPsp32. WinXP detected the new USB device and loaded the BT drivers to support it:
Microsoft Bluetooth Enumerator v5.1.2600.2180
USB Bluetooth Wireless Adapter v5.1.2600.2180
2) I inserted a second BT/USB dongle in a USB port of my laptop computer running WinXPsp2. WinXP detected the new USB device and loaded the same version drivers as the desktop system.
3) Opened LV8.0 on the desktop system and ran Simple Bluetooth Server.vi. VI correctly lists the "My Bluetooth Name" as the name of the my desktop computer, displays the hardware address of the BT dongle and the channel number. (Is this where your app fails?)
4) Opened LV8.0 on the laptop system and ran Simple Bluetooth Client.vi. I entered the hardware address of the desktop BT adapter and the correct channel number. By the time I finished entering the hex values, 30 sec had passed and the Simple BT Server.vi had to be restarted, as it had timed out and stopped.
5) After restarting the Server.vi, I started the Client.vi and --- magic.... it works as advertised.
Is this similar to your experience?
-Bill =]Message Edited by qswitch on 09-28-2006 11:14 AM

Similar Messages

  • Plc data logging using Labview

    I am trying to communicate with PACSystem RX3i using Labview in order to log
    data from PLC. Is this possible?
    If it is possible, how could I implement it?

    Hi mber,
    Thanks for posting at NI forums. I looked into the supported device & driver plug-in list for NI OPC Servers, and the PACSystem RX3i is supported. To configure the OPC server to communicate with this PLC, you can follow the steps given in this article. In the LabVIEW side, you can configure an I/O server as an OPC client to communicate with the server.
    In order to configure the I/O server you will require the Datalogging and Supervisory Control (DSC) module installed for LabVIEW. In the article I just referenced, it shows how to configure bound variables to communicate with the PLC tags to retrieve data. The DSC module installs a Citadel database, so when you are configuring the bound variable, you can choose to log data depending on events, alarms or a historical log. This article describes the data logging options using the DSC module. You can also choose to read and display the data in a VI by dragging the bound variables to the block diagram and then log to a file using logging functions.
    Best Regards,
    Alina M
    Applications Engineering
    National Instruments

  • Acquire, analyze and data log using statemachine technique

    I am learning state machine technique to develop code for data acquisitoin and logging. I think I have written the correct code for this purpose but I couldn't get the following done:
    All warning values(>0) should be displayed on 1d array on front panel. The code is writing only the first value.
    All values should be logged using write to spreadsheet pallete and be written on excel sheet. Code is only writing first value.
    I wonder what's the mistake. The code I've created is attached, I'd appreciate if anyone correct the code and post in reply.
    Many thanks in advance for help
    Cheers
    Solved!
    Go to Solution.
    Attachments:
    full project with acquire,analyse and logging states.vi ‏94 KB

    Hi kwaris,
    "what's the mistake": Well, you use a lot of unneccessary conversions. Really a lot of...
    Why do you convert a scalar to dynamic, then convert to array, then index an element of the array? Why do you convert to dynamic when all you needis a simple "build array" node?
    Ok, I included a shift register to store your array. It's only a simple "how to", but not the best solution for all cases. But it should give you a hint...
    Best regards,
    GerdW
    CLAD, using 2009SP1 + LV2011SP1 + LV2014SP1 on WinXP+Win7+cRIO
    Kudos are welcome
    Attachments:
    full project with acquire,analyse and logging states.vi ‏12 KB

  • DS 4.2 get ECC CDHDR deltas in ABAP data flow using last run log table

    I have a DS 4.2 batch job where I'm trying to get ECC CDHDR deltas inside an ABAP data flow.  My SQL Server log table has an ECC CDHDR last_run_date_time (e.g. '6/6/2014 10:10:00') where I select it at the start of the DS 4.2 batch job run and then update it to the last run date/time at the end of the DS 4.2 batch job run.
    The problem is that CDHDR has the date (UDATE) and time (UTIME) in separate fields and inside an ABAP data flow there are limited DS functions.  For example, outside of the ABAP data flow I could use the DS function concat_date_time for UDATE and UTIME so that I could have a where clause of 'concat
    _date_time(UDATE, UTIME) > last_run_date_time and concat_date_time(UDATE, UTIME) <= current_run_date_time'.  However, inside the ABAP data flow the DS function concat_date_time is not available.  Is there some way to concatenate UDATE + UTIME inside an ABAP data flow?
    Any help is appreciated.
    Thanks,
    Brad

    Michael,
    I'm trying to concatenate date and time and here's my ABAP data flow where clause:
    CDHDR.OBJECTCLAS in ('DEBI', 'KRED', 'MATERIAL')
    and ((CDHDR.UDATE || ' ' || CDHDR.UTIME) > $CDHDR_Last_Run_Date_Time)
    and ((CDHDR.UDATE || ' ' || CDHDR.UTIME) <= $Run_Date_Time)
    Here are DS print statements showing my global variable values:
    $Run_Date_Time is 2014.06.09 14:14:35
    $CDHDR_Last_Run_Date_Time is 1900.01.01 00:00:01
    The issue is I just created a CDHDR record with a UDATE of '06/09/2014' and UTIME of '10:48:27' and it's not being pulled in the ABAP data flow.  Here's selected contents of the generated ABAP file (*.aba):
    PARAMETER $PARAM1 TYPE D.
    PARAMETER $PARAM2 TYPE D.
    concatenate CDHDR-UDATE ' ' into ALTMP1.
    concatenate ALTMP1 CDHDR-UTIME into ALTMP2.
    concatenate CDHDR-UDATE ' ' into ALTMP3.
    concatenate ALTMP3 CDHDR-UTIME into ALTMP4.
    IF ( ( ALTMP4 <= $PARAM2 )
    AND ( ALTMP2 > $PARAM1 ) ).
    So $PARAM1 corresponds to $CDHDR_Last_Run_Date_Time ('1900.01.01 00:00:01') and $PARAM2 corresponds to $Run_Date_Time ('2014.06.09 14:14:35').  But from my understanding ABAP data type D is for date only (YYYYMMDD) and doesn't include time, so is my time somehow being defaulted to '00:00:00' when it gets to DS?  I ask this as a CDHDR record I created on 6/6 wasn't pulled during my 6/6 testing but this 6/6 CDHDR record was pulled today.
    I can get  last_run_date_time and current_run_date_time into separate date and time fields but I'm not sure how to build the where clause using separate date and time fields.  Do you have any recommendations or is there a better way for me to pull CDHDR deltas in an ABAP data flow using something different than a last run log table?
    Thanks,
    Brad

  • How to data log graphs using front panel data logging?

    Hello I have a VI that collects data from DAQmx thermocouple readings and graphs the temperature vs time using a while loop to collect data and graph. I have 9 control operators that define the correction factor of the thermocouples.
    I want to create a datalogging using the option under Operations>Data Logging
    When I retrieve the data the only information that is present are the control operators correction factors that I defined. The graphed data that was created is not retrieved.
    Is there a solution to show the graphed data plots that were created on the front panel? They remain unchanged from the last run of the VI or blank if I open the VI without having ran the program.
    Thank you.

    This is expected for the Data Logging in LabVIEW. If you want to record the signla data, use the Write to Measurement File Express VI.  Here's a link with a walk-through:
    http://www.ni.com/academic/students/learn-daq/data-logging/
    The Data Logging from the Operate Menu is for recording front panel control(s), as you have observed.
    Mark P.
    Applications Engineer
    National Instruments
    www.ni.com/support

  • Rapid disk use with DSC data logging

    I recently installed Labview 8.2.1 with DSC module.  I have tried a few shared variable projects with data logging and now I find my hard drive space is being consumed very rapidly.  The variable manager indicates that only a few shared variables are left from my projects.  They would produce very little data in the Citadel database.  Certainly not enough to consume 100 meg + of disk per day.  I am obviously missing something that would turn off some hidden logging process.  Any suggestions?

    Thanks for the input.  I did have another look at historical data and there are not large traces.  I used the Variable manager to stop all processes and I stopped the shared variable engine.  The leak continues.  The only way I can stop it is to use the Windows Task Manager to shut down the Citadel5 process.  I have searched the computer and cannot find any large database files that would account for the disk shrink.  I have also uninstalled and re-installed the DSC module.  No success.  
    If I knew where the data was located I might be able to determine the source.  The National Instruments directory is not growing even though all the processes had their database pointed to a directory inside the NI directory!
    Len

  • ISSUE:regarding production version tab of MRP4 of  data transfer using BD10

    Hi all,
    This is regarding production version tab of MRP4 of  data transfer using ALE idoc (BD10).
    When i transfer the data using BD10 the production version is received at receiving end through segment
    but not created in MRP4 view tab.An error is coming as-
    "You wanted to maintain the master record of the material AB_06.04.09(2). However, it is already being processed by the user EBGABAP and is therefore locked."
    I logged in as-EBGABAP
    Please help me resolve the issue.
    Thanks
    Edited by: sanu debu on May 6, 2009 11:08 AM
    Edited by: sanu debu on May 6, 2009 11:09 AM
    Edited by: sanu debu on May 6, 2009 11:11 AM

    >
    sanu debu wrote:
    > Hi,
    > I have to upload production version tab data of  MRP4 view(MM01).Please suggest a function  moduleor bapi  for the same.
    >
    > Thanks.
    >
    > Edited by: sanu debu on May 6, 2009 3:24 PM
    BAPI_MATERIAL_SAVEDATA can be used, populate the respective fields in input parameter PLANTDATA

  • Extracting SQL statement from a Webi document's data provider using SDK.

    Hi all,
    Is it possible to extract the SQL statement from an existing Webi document's data provider using BO SDK?  I've searched through the class library but haven't found any information on this yet.  If you have done it, could you provide some guidance.  Many thanks.

    I found the following Java code that might be of some help to you. I realize you are using .NET but this might push you down the right path.
    The trick here is to use the Report Engine SDK to get the DataProvider of the DocumentInstance. Then, look at the SQLDataProvider to get your SQLContainer.
    My apologies for the poor formatting. This didn't copy and paste over to the forums very well. I've cleaned up as much as I could.
    <%@ page import="com.crystaldecisions.sdk.framework.*" %>
    <%@ page import="com.crystaldecisions.sdk.exception.SDKException" %>
    <%@ page import="com.crystaldecisions.sdk.occa.infostore.*" %>
    <%@ page import="com.businessobjects.rebean.wi.*" %>
    <%
    boolean loginSuccessful = false;
    IEnterpriseSession oEnterpriseSession = null;
    String username = "username";
    String password = "password";
    String cmsname  = "cms_name";
    String authenticationType = "secEnterprise";
    try
    //Log in. oEnterpriseSession = CrystalEnterprise.getSessionMgr().logon( username, password, cmsname, authenticationType);
    if (oEnterpriseSession == null)
    out.print("<FONT COLOR=RED><B>Unable to login.</B></FONT>");
    else
    {  loginSuccessful = true;
    catch (SDKException sdkEx)
    { out.print("<FONT COLOR=RED><B>ERROR ENCOUNTERED</B><BR>" + sdkEx + "</FONT>");}
    if (loginSuccessful) { IInfoObject oInfoObject = null;
    String docname = "WebI document name";
    //Grab the InfoStore from the httpsession IInfoStore oInfoStore = (IInfoStore) oEnterpriseSession.getService("", "InfoStore");  //Query for the report object in the CMS.  See the Developer Reference guide for more information the query language.   String query = "SELECT TOP 1 * " +        "FROM CI_INFOOBJECTS " +        "WHERE SI_INSTANCE = 0 And SI_Kind = 'Webi' " +        "AND SI_NAME='" + docname + "'";
    IInfoObjects oInfoObjects = (IInfoObjects) oInfoStore.query(query);
    if (oInfoObjects.size() > 0)
    //Retrieve the latest instance of the report  oInfoObject = (IInfoObject) oInfoObjects.get(0);
      // Initialize the Report Engine  ReportEngines oReportEngines = (ReportEngines)
    oEnterpriseSession.getService("ReportEngines"); 
    ReportEngine oReportEngine = (ReportEngine) oReportEngines.getService(ReportEngines.ReportEngineType.WI_REPORT_ENGINE);
      // Openning the document  DocumentInstance oDocumentInstance = oReportEngine.openDocument(oInfoObject.getID());   
    DataProvider oDataProvider = null; 
    SQLDataProvider oSQLDataProvider = null; 
    SQLContainer oSQLContainer_root = null; 
    SQLNode oSQLNode = null;
    SQLSelectStatement oSQLSelectStatement = null; 
    String sqlStatement = null;
      out.print("<TABLE BORDER=1>");
    for (int i=0; i<oDocumentInstance.getDataProviders().getCount(); i++)
    oDataProvider = oDocumentInstance.getDataProviders().getItem(i);
      out.print("<TR><TD COLSPAN=2 BGCOLOR=KHAKI>Data Provider Name: " + oDataProvider.getName() + "</TD></TR>");
       if (oDataProvider instanceof SQLDataProvider)
    oSQLDataProvider = (SQLDataProvider) oDataProvider;
        oSQLContainer_root = oSQLDataProvider.getSQLContainer();
        if (oSQLContainer_root != null)
    for (int j=0; j<oSQLContainer_root.getChildCount(); j++)
    oSQLNode = (SQLNode) oSQLContainer_root.getChildAt(j);
          oSQLSelectStatement = (SQLSelectStatement) oSQLNode;             
    sqlStatement = oSQLSelectStatement.getSQL();     
    out.print("<TR><TD>" + (j+1) + "</TD><TD>" + sqlStatement + "</TD></TR>");   
    else
    out.print("<TR><TD COLSPAN=2>Data Provider is not a SQLDataProvider.  SQL Statement can not be retrieved.</TD></TR>");   }  }  out.print("</TABLE>");
      oDocumentInstance.closeDocument(); }
    oEnterpriseSession.logoff();}%>

  • How can I perform data logging for a specific time??

    hello everyone,
    I am quite new in labview and I have a basic question regarding data logging. Currently I am using a cRIO9074 and doing some data logging for my test. The data logging it self works ok so far.
    But my problem is I would like to write my datas in a text file either for a specific time interval (ex)10 seconds) or for a specific amounts of data (ex)500 Samples). Can anyone give me some help regarding my problem?? Attached you can find my RT.vi 
    I would appreciate for anyhelp!
    Regards
    Yun 
    Attachments:
    BP250 Encoder Position & Velocity (Host).vi ‏92 KB

    Run your loggging program for that time. When your program terminates then it will write all the logged data so far in text file.
    Kudos are always welcome if you got solution to some extent.
    I need my difficulties because they are necessary to enjoy my success.
    --Ranjeet

  • How do I control a data log session with period and sample time?

    I need a data logging system where the operator can select 2 logging parameters: Log Period and Sample Time. I also need a START and STOP button to control the logging session. For example, set the log period for 1 hour and the sampling time for 1 second. (I may be using the wrong jargon here.) In this case when the START button is clicked, the system starts logging for 1 second. An hour later, it logs data for another second, and so on until the operator clicks the STOP button. (I will also include a time limit so the logging session will automatically stop after a certain amount of time has elapsed.)
    It’s important that when the STOP button is clicked, that the system promptly stops logging. I cannot have the operator wait for up to an hour.
    Note that a logging session could last for several days. The application here involves a ship towing a barge at sea where they want to monitor and data log tow line tension. While the system is logging, I need the graph X-axis (autoscaled) to show the date and time. (I’m having trouble getting the graph to show the correct date and time.) For this application, I also need the system to promptly start data logging at a continuous high rate during alarm conditions.
    Of course I need to archive the data and retrieve it later for analysis. I think this part I can handle.
    Please make a recommendation for program control and provide sample code if you can. It’s the program control concepts that I think I mostly need help here. I also wish to use the Strip Chart Update Mode so the operator can easily view the entire logging session.
    DAQ Hardware: Not Selected Yet
    LabVIEW Version: 6.1 (Feel free to recommend a v7 solution because I need to soon get it anyway.)
    Operating System: Win 2000
    In summary:
    How do I control a graphing (data log) session for both period and sample time?
    How do I stop the session without having to wait for the period to end?
    How do I automatically interrupt and control a session during alarm conditions?
    Does it make a difference if there is more than one graph (or chart) involved where there are variable sample rates?
    Thanks,
    Dave

    Hello Dave,
    Sounds like you have quite the system to set up here. It doesn�t look like you are doing anything terribly complicated. You should be able to modify different examples for the different parts of your application. Examples are always the best place to start.
    For analog input, the �Cont Acq&Chart (buffered).vi� example is a great place to start. You can set the scan rate (scans/second) and how many different input channels you want to acquire. This example has its own stop button; it should be a simple matter to add a manual start button. To manually set how long the application runs, you could add a 100 ms delay to each iteration of the while loop (recommended anyway to allow processor to multi-task) and add a control that sets the number
    of iterations of the while loop.
    For logging data, a great example is the �Cont Acq to File (binary).vi� example.
    For different sample rate for different input lines, you could use two parallel loops both running the first example mentioned above. The data would not be able to be displayed on the same graph, however.
    If you have more specific questions about any of the different parts of your application, let me know and I�ll b happy to look further into it.
    Have a nice day!
    Robert M
    Applications Engineer
    National Instruments
    Robert Mortensen
    Software Engineer
    National Instruments

  • With journaling, I have found that my computer is saving a large amount of data, logs of all the changes I make to files; how can I clean up these logs?

    With journaling, I have found that my computer is saving a large amount of data, logs of all the changes I make to files; how can I clean up these logs?
    For example, in Notes, I have written three notes; however if I click on 'All On My Mac' on the side bar, I see about 10 different versions of each note I make, it saves a version every time I add or delete a sentence.
    I also noticed, that when I write an email, Mail saves about 10 or more draft versions before the final is sent.
    I understand that all this journaling provides a level of security, and prevents data lost; but I was wondering, is there a function to clean up journal logs once in a while?
    Thanks
    Roz

    Are you using Microsoft word?  Microsoft thinks the users are idiots. They put up a lot of pointless messages that annoy & worry users.  I have seen this message from Microsoft word.  It's annoying.
    As BDaqua points out...
    When you copy information via edit > copy,  command + c, edit > cut, or command +x, you place the information on the clipboard. When you paste information, edit > paste or command + v, you copy information from the clipboard to your data file.
    If you edit > cut or command + x and you do not paste the information and you quite Word, you could be loosing information.  Microsoft is very worried about this. When you quite Word, Microsoft checks if there is information on the clipboard & if so, Microsoft puts out this message.
    You should be saving your work more than once a day. I'd save every 5 minutes.  command + s does a save.
    Robert

  • Issue found in EHS in using specification data import using  process

    Dear EHS community
    Now using EHS classic for a long time a issue has been detected in EHS standard import. During maintenance of EHS data normally using CG02 the system is always using the default "Data origin" specified in customizing to be stored in EHS tables (e.g. ESTRH, ESTRI etc.). In standard process to import specification data one can define a different "Data origin". Now we are using an import file with default" data origin and executed the import. Now a strange effect has been detected (and not always) for update of identifiers. For the import purpose you must nominate at least one identifier. If the identifier is found then normally no update happens but only the value assignment data is inserted (or updated). If the identifier is not found it get be inserted on spec level in ESTRI. Now during the update the "Data origin" of the identifier present in the system (and which matched to identifier on file level) was changed but not the identifier as such. Any data record on value assignment level received the defautl data origin. Actually there is no explanation for this behaviour. If Default "Data origin" would be "SAP" (as the term) this value has been change to "Space". Any explanation of this effect is appreciated (or and idea regarding that).
    C.B.
    PS: analysis of change logs in EHS etc. executed so far clearly indictae that an "Update" happened on the identifier; but only the field SRSID is effected; EHS import is quite old and therefore very stable;
    PPS: I found a thread taking about the import file:
    spec import_inheritance data
    Example shown thre is like:
    +BS
    +BV   $ESTRH
    SRSID                          EH&S
    SUBID                          000000385000
    SUBCAT                         REAL_SUB
    AUTHGRP                        EHS_PS
    +EV
    +BV   $ESTRI
    SRSID                          EH&S
    IDTYPE                         NAM
    IDCAT                          EHS_OLD
    IDENT                          XY0002
    ORD                            0001
    +EV
    +BV   SAP_EHS_1013_001
    $ESTVA-SRSID                   EH&S
    SAP_EHS_1013_001_VALUE         N09.00101280
    +EV
    If you compare SAP helpt normally only at the"begining pof the file you will find "SRSID" Here this field is nominated often. On Level of ESTRH as well as ESTRI.
    PPS: e.g. refer to: TCG56 EHS: Data Origin - SAP Table - ABAP

    Dear Ralph
    first thanks for feedback. Regarding content provider: we need to check that on deeper level
    Regarding import may be my explanation was not good enough.
    Imagine this case:
    you have a specification in the system there you would like to add e.g. denisity data.. To do so you need at least one identifier which you must nominate during the import. As long as this identifier is "identical" in system and in the file this identifier should not be "changed/effected" etc. and only the additional data should be loaded, This was the process we used. Now we detected that this seems not to be the real effect. The identifier as part of the file is "updated" in EH&S. In the example above somebody used this logic:
    +BV   $ESTRI
    SRSID                          EH&S
    IDTYPE                         NAM
    IDCAT                          EHS_OLD
    IDENT                          XY0002
    ORD                            0001
    +EV
    Nearly the same is used in our process. Only difference ist, that we do not define the "SRSID". The same is true for any other data in the file. SRSID is nether specified.
    What is happing now:
    e.g. the "density" data is added with SRSID "EH&S". This effect is "normal". As by default SRSID should be EH&S (as this is defined as such in customizing) and because of the fact that at the top the ID is es well "EH&S". In the system we have the "XY0002" having SRSID EH&S. By using now this upload approach the only difference afterwards is tht kin th systeM; XY0002 does get "blank" as SRSID (and there is no data origin "blank" defined. Up to today my unertsanding was clearly: no update should happen in the identifier. This seems not to be the case. Is my understanding here different? Or is this SRSID in the load file is really mandatory in ESTRI level to avoid this effect. I hope that you can provide some feedback regarding this.
    C.B.
    PS: referring to: Example: Transfer File for Specifications - Basic Data and Tools (EHS-BD) - SAP Library
    The header of import file should look like:
    Comment
    +C
    Administrative section
    Character standard
    +SC
    ISO-R/3
    Identification (database name)
    +ID
    IUCLID
    Format version
    +V
    2.21
    Export date
    +D
    19960304
    Key date for export
    +VD
    19960304
    Set languages for export
    +SL
    E
    Date format
    +DF
    DD.MM.YYYY
    IN our case +ID = EH&S (as this is the value in export file)
    IN this example this additional one is shown:
    Begin table
    +BV
    $ESTRI
      Table field
    IDTYPE
    NAM
      Table field
    IDCAT
    IUPAC
      Table field
    IDENT
    anisole
      Table field
    LANGU
    E
      Table field
    OWNID
    ID1
    Therefore no SRSID is specified. And this is the data in our file (on high level) and the "only" change is that the identifer get "deleted" the SRSID

  • Help please: DAQWare or data logging software (multichannel data capture) questions

    I am writing regarding DAQWare, data logging software, and the PC-LPM-16
    board, which a group of us a trying to come to grips with. This is rather
    lengthy, I'm afraid...
    We are a group of college students in a computer programmer/analyst
    program. We have a course that involves doing a project to go through the
    software development lifecycle (or as much of it as can be done in a term).
    For our project we teamed up with a group in the industrial hygiene
    program; they have a computer (standalone) with a PC-LPM-16 (PnP) DAQ board
    and the NI-DAQ software (all of this has been around for a few years, but
    has not been used much), and our goal was to capture the data they are
    measuring (atmospheric pollutants) to an Microsoft Access database.
    We are not familiar with DAQ systems, but were shown the software they had
    installed. This included DAQWare (appears to be version 2.0). We thought
    we could use DAQWare to log data to a text file, which we then would import
    into the Access database. We came up with VBA code to parse the ASCII text
    file created, but have run into one problem: DAQWare does not timestamp
    each sample, but only puts a single timestamp in the header information.
    Now we could extrapolate the time for each line based on the sample rate,
    except that sometimes it seems to put in 8 values per sample, sometimes 12
    values. For instance, in the sample of a text file below, the capture was
    run for 1 minute at a sample rate of 60, and there were 480 measurements in
    the file, or 8 lines of measurements per sample. (At a sample rate of 100,
    we got 800 measurements in the file etc.) This was two weeks ago. Today,
    we returned to capture more data, and DAQWare was logging 12 measurements
    per sample (so at a sample rate of 100, we ended up with 1200 lines of
    measurements). We do not know where this factor of 8 or 12 is coming from
    or how it got changed in the two weeks between captures. We would like to
    be able to have them just run DAQWare, for days on end if they wanted,
    capture the data to file, then use our code to parse the text file into the
    database, but if we cannot put timestamps on the readings, there seems no
    point.
    *** Example of a capture below... ***
    DAQWare File
    Version: 2.0
    Number of Lines in Header: 10
    Comment:
    Date: 10-18-1999 Time: 10:01:39
    Sample Rate: 60.0
    Scan Rate: 0.0
    File Format: 1
    Samples per Channel: 480.0
    Number of Channels: 5
    O3 SO2 NOX NO2 NO
    4 3 2 1 0
    1 1 1 1 1
    0.537 0.276 0.259 -0.017 0.015
    0.537 0.273 0.261 -0.022 0.012
    0.537 0.276 0.259 -0.037 0.012
    0.537 0.276 0.264 -0.007 0.015
    *** and so on... ***
    Regarding this attempt to use DAQWare:
    1. Can you tell us where this factor (8 or 12 or however many more readings
    than we would expect) means, where it is read from or set?
    2. Is there an updated version of DAQWare that we should download that
    would be better?
    3. Is there some other existing software that we can download to use to
    perform basic logging of multichannel analog input to a text file (with
    timestamp)?
    I realise that the software comes with tools (DLL functions) that one can
    use to create custom software to interface with the DAQ board, but in the
    limited amount of time we have we are not sure we can learn enough about
    these functions to write our own datalogging software.
    However, to see if I could grasp the basics, I have looked through the
    examples for Visual Basic and think that a program would run something like
    this:
    1. at the beginning, use LPM16_Calibrate to reset the board
    2. configure acquisition with DAQ_Config
    3. start acquisition with Lab_ISCAN_Start
    4. loop to check is acquisition done with Lab_ISCAN_Check
    5. once sampling done, write data out to file
    6. loop back up to 3 to get next lot of samples
    Unfortunately, a sample app using Access and VBA I created came back with
    an error telling me it could not find nidaq.dll (even though this is in the
    Windows system directory); probably it is not a good idea to try to do this
    in Access in the first place. I am going to have a go using Delphi, which
    I know a bit better.
    One thing I am unsure of is what the numbers will be coming out of
    Lab_ISCAN; the reference file seems to indicate they will be binary
    readings, not voltages, and I wonder how to correlate these with instrument
    measurements (the instrumentation people we going to set it up so that
    1mV=1ppm).
    Any help you can give to point us in the right direction would be
    appreciated.
    Sincerely, Marc Bell ([email protected])

    I am writing regarding DAQWare, data logging software, and the PC-LPM-16
    board, which a group of us a trying to come to grips with. This is rather
    lengthy, I'm afraid...
    We are a group of college students in a computer programmer/analyst
    program. We have a course that involves doing a project to go through the
    software development lifecycle (or as much of it as can be done in a term).
    For our project we teamed up with a group in the industrial hygiene
    program; they have a computer (standalone) with a PC-LPM-16 (PnP) DAQ board
    and the NI-DAQ software (all of this has been around for a few years, but
    has not been used much), and our goal was to capture the data they are
    measuring (atmospheric pollutants) to an Microsoft Access database.
    We are not familiar with DAQ systems, but were shown the software they had
    installed. This included DAQWare (appears to be version 2.0). We thought
    we could use DAQWare to log data to a text file, which we then would import
    into the Access database. We came up with VBA code to parse the ASCII text
    file created, but have run into one problem: DAQWare does not timestamp
    each sample, but only puts a single timestamp in the header information.
    Now we could extrapolate the time for each line based on the sample rate,
    except that sometimes it seems to put in 8 values per sample, sometimes 12
    values. For instance, in the sample of a text file below, the capture was
    run for 1 minute at a sample rate of 60, and there were 480 measurements in
    the file, or 8 lines of measurements per sample. (At a sample rate of 100,
    we got 800 measurements in the file etc.) This was two weeks ago. Today,
    we returned to capture more data, and DAQWare was logging 12 measurements
    per sample (so at a sample rate of 100, we ended up with 1200 lines of
    measurements). We do not know where this factor of 8 or 12 is coming from
    or how it got changed in the two weeks between captures. We would like to
    be able to have them just run DAQWare, for days on end if they wanted,
    capture the data to file, then use our code to parse the text file into the
    database, but if we cannot put timestamps on the readings, there seems no
    point.
    *** Example of a capture below... ***
    DAQWare File
    Version: 2.0
    Number of Lines in Header: 10
    Comment:
    Date: 10-18-1999 Time: 10:01:39
    Sample Rate: 60.0
    Scan Rate: 0.0
    File Format: 1
    Samples per Channel: 480.0
    Number of Channels: 5
    O3 SO2 NOX NO2 NO
    4 3 2 1 0
    1 1 1 1 1
    0.537 0.276 0.259 -0.017 0.015
    0.537 0.273 0.261 -0.022 0.012
    0.537 0.276 0.259 -0.037 0.012
    0.537 0.276 0.264 -0.007 0.015
    *** and so on... ***
    Regarding this attempt to use DAQWare:
    1. Can you tell us where this factor (8 or 12 or however many more readings
    than we would expect) means, where it is read from or set?
    2. Is there an updated version of DAQWare that we should download that
    would be better?
    3. Is there some other existing software that we can download to use to
    perform basic logging of multichannel analog input to a text file (with
    timestamp)?
    I realise that the software comes with tools (DLL functions) that one can
    use to create custom software to interface with the DAQ board, but in the
    limited amount of time we have we are not sure we can learn enough about
    these functions to write our own datalogging software.
    However, to see if I could grasp the basics, I have looked through the
    examples for Visual Basic and think that a program would run something like
    this:
    1. at the beginning, use LPM16_Calibrate to reset the board
    2. configure acquisition with DAQ_Config
    3. start acquisition with Lab_ISCAN_Start
    4. loop to check is acquisition done with Lab_ISCAN_Check
    5. once sampling done, write data out to file
    6. loop back up to 3 to get next lot of samples
    Unfortunately, a sample app using Access and VBA I created came back with
    an error telling me it could not find nidaq.dll (even though this is in the
    Windows system directory); probably it is not a good idea to try to do this
    in Access in the first place. I am going to have a go using Delphi, which
    I know a bit better.
    One thing I am unsure of is what the numbers will be coming out of
    Lab_ISCAN; the reference file seems to indicate they will be binary
    readings, not voltages, and I wonder how to correlate these with instrument
    measurements (the instrumentation people we going to set it up so that
    1mV=1ppm).
    Any help you can give to point us in the right direction would be
    appreciated.
    Sincerely, Marc Bell ([email protected])

  • Data logging optimization function in LabVIEW and SignalExpress

    Hello, everyone!  I and a colleague have assembled the attached VI,
    which is used to control a plant growth chamber, and I could use some advice. 
    I would like to add data logging capabilities, ideally to note (1)
    environmental conditions, such as temp and CO2 level and (2) when certain
    situations occur, such as high CO2 or low pressure (which results in a change
    to the “Case Structure for CO2 and Pressure”).
    The chamber runs for weeks at a time, though, so I have a constraint.
    Instead of continuously logging all the environmental data (which would yield
    giant file sizes), I want to be able to take “snapshots”, say every 5 minutes,
    so I can examine the chamber’s condition over a long period of time.
    I attempted to use the Time Delay function to control the Write Measurements
    function (as in LabVIEW’s Cycle Analysis example, which I’ve also attached),
    but Time Delay halted my entire program. 
    I also tried the Wait (ms) function, but had no success.
    I recently discovered the powerful data review and reporting tools that
    SignalExpress has to offer.  Ideally, I
    would like to use a SignalExpress express VI to record the measurements instead of the more primative Write Measurements, but, I’m not sure how to implement this in my VI.  For example, would I make a data acquisition VI in SignalExpress for all my sensors, export to LabVIEW, and use that to replace the current DAQ structure in my VI?  (Because the chamber VI is interactive, I'm pretty sure that porting the LabVIEW VI to SignalExpress wouldn't work out well.)  And I’m
    still not sure how to control how often it records measurements.
    Any advice would be much appreciated!
    Attachments:
    Chamber VI.vi ‏789 KB
    Cycle Analysis.vi ‏300 KB

    Hi, Sarah.  Yes I definitely plan to use LabVIEW for the majority of my data acquisition and logging.  I apologize, I think I probably should have broken my post into two separate posts for clarity, as my major concern is being able to set how often LabVIEW logs the environmental data (viewing the logged data in SignalExpress would be nice, but not a requirement).
    As I explained before, I'd really like to be able to set up a system that records the environmental condtions at regular time intervals, say every 30 minutes, or if my "Case Structure for CO2 and Pressure" status changes (such as from "Normal Conditions" to "High CO2".
    As an example, attached is a boolean structure (TimedWritedMeasurements.vi) that I came up with to record a simulated signal every 5 seconds (5000 ms).  Once I can get it to work, I would connect the boolean to my environmental sensor outputs and CO2 and Pressure case structure in my Chamber VI.vi that I attached in my first post.
    My problem is I can't seem to get the Trigger and Gate function to work the way I'd like.  I'd like the Wait (ms) function to trigger the Trigger and Gate function every 5 seconds to change the case structure in the TimedWriteMeasurements.vi to "True", thereby causing the Write Measurement function to record the environmental conditions.  Then I'd like the case structure reset to false, until it's triggered to true in another five seconds.  Can anyone point me in the right direction?
    Attachments:
    TimedWriteMeasurements.vi ‏84 KB

  • Data plan used if MiFi connected but not actively being used?

    If computer is on overnight but not in use, meaning MiFi connected, with that use up my data plan? Or does it use up with "active" use only? I'm sharing internet with other people so I'm not sure my data plan is being used on purpose or just getting eaten up while we are all asleep? I'm already maxed with 3 weeks to go!!!

    As long as you have devices connected to the MiFi there is always a risk that those devices will consume data behind the scenes without your knowledge.  This is how network connected devices work, they are constantly "in use" while connected.  Computers, smart phones and any kind of internet connected devices are constantly backing up content, looking for updates and talking to other network devices.  It has nothing to do with your active participation and all internet communication is held against you on your data plan.
    The only way to completely stop data usage is to turn off your MiFi when it is not needed.  This is a common safe practice that you should start using when on a metered data plan.  MiFis are not intended to be left on 24x7 anyways, frequent reboots will help ensure the device stays in proper working order and increase the longevity of its life.
    Verizon does provide everyone with a data log.  The data log will provide you with time stamps and the amounts that are consumed at those times.  It should be easy to see when your biggest spikes of consumption are.  Isolate what devices are online during those times and what may have been going on and you can start to build a picture of where the problem lies.
    To truly see the network traffic in terms of specific applications, web services or devices you need to find and install network monitors.  Network monitors come in a variety of applications so you need to find and build a monitoring system that fits your environment.  Considering you have other people connected to your MiFi it would not be uncommon for each of you to install a separate monitor specific to those machines.
    Let me know if you have any other questions.

Maybe you are looking for

  • New ipod won't show on itunes but old ipod will

    I bought a brand new 5th gen ipod and installed the software and it only installed the ipod updater, so as i already have a 3rd gen ipod and itunes on the computer i uninstalled all ipod and itunes software that had previously been downloaded and sta

  • PLEASE HELP ME!! Ipod not recognised in i-tunes or by computer

    My pod used to be recognised!! No probs, but now for some reason there is no "device" tab at the side and no recognition of it anywhere. But it still charges when plugged in?! I have tried the 5 Rs (as much as I can- can't do the last one because the

  • No v9 properties toolbar access except in last document opened

    Has anyone else noticed this in v9.0 or 9.1 of acrobat? You want to change the properties of an object such as the text format in a comment/markup text box and, even though you can select the text, "No current selection" is reported where the propert

  • Problems Migrating from iPhoto 6 to 11

    I'm moving from an older Macbook Pro to a brand new one, and used Migration Assistant to help get everything moved over. iPhoto 11 on the new machine found most of my photos but there were some anomalies. I'm wondering if anybody can shed any light o

  • Exception in XMLIsValid function

    Hi anybody can you help me in providing the actual error whic i am geetiing while validating the xml with xsd by using XMLIsValid  function