Beginner question on labview and transducer

i'm trying to understand what a transducer does and what labview can do
to it.
are these assumption correct?
1) with labview, i can change the frequency of sound waves emitted from
a ultrasonic transducer. if i want maximum amplitude of the sound wave
(loudness), i would set the transducer to its internal resonant
frequency (1 mHz) with labview. if i set the transducer to any other
frequency with labview, i would not get max amplitude.
2) input power of the transducer is 100 W which is the max amount of
power i can supply to the transducer. the more power i supply the
higher the amplitude of the sound wave (with 100 W the maximum power the
transducer can handle) regardless of the resonant frequency. Resonan
t
frequncy and the amount of input combined both effect the amplitude of
the sound waves.
3) can labview change the resonant frequcy and amount of input power to
the transducer by itself by simply clicking a switch or turning a dial
on the labview front panel?
4) any good books to help me write a labview GUI for a ultrasonic
transducer to measure the distance traveled and the time taken of a
reflected sound wave in water?
5) what is beamwidth (3 degree) and depth rating (100 m) of a
transducer?
thanks

[email protected] wrote:
> i'm trying to understand what a transducer does and what labview can do
> to it.
A transducer by definition changes one form of energy into another.
Typically it turns some sort of input to a voltage (a microphone
changes sound waves to a voltage) or it changes a voltage
to some form of output (a speaker changes voltage to sound waves).
I am not familiar with the transducer you are using but I will try to work
using
some assumptions and my used brain.
>
> are these assumption correct?
> 1) with labview, i can change the frequency of sound waves emitted from
> a ultrasonic transducer. if i want maximum amplitude of the sound wave
> (loudness), i would set the transducer to its internal resonant
> frequency (1 mHz) with labview. if i set the transducer to any other
> frequency with labview, i would not get max amplitude.
Can not say since I have no idea how your transducer works.
But this sounds reasonable
>
>
> 2) input power of the transducer is 100 W which is the max amount of
> power i can supply to the transducer. the more power i supply the
> higher the amplitude of the sound wave (with 100 W the maximum power the
>
> transducer can handle) regardless of the resonant frequency. Resonant
> frequncy and the amount of input combined both effect the amplitude of
> the sound waves.
Again dont know but this sounds reasonable.
>
>
> 3) can labview change the resonant frequcy and amount of input power to
> the transducer by itself by simply clicking a switch or turning a dial
> on the labview front panel?
Not by itself since LabView is software, you will need some sort of hardware
to connect to the transducer and Labview to control that hardware
>
>
> 4) any good books to help me write a labview GUI for a ultrasonic
> transducer to measure the distance traveled and the time taken of a
> reflected sound wave in water?
>
> 5) what is beamwidth (3 degree) and depth rating (100 m) of a
> transducer?
> thanks
Consider a flashlight. Turn it on and point it at the wall. The size of the
spot it makes
is its beamwidth (or maybe half of that). It is measured in degrees because
it is defined
as the angle between the center and the edge.
I am not sure about the depth rating. Since you are talking about water this
may be
how far down you can put it.
It sounds to me that you will need a few pieces of hardware to accomplish
your task.
1) Some sort of Digital Acquisition board (DAQ) NI makes several.
2) Some sort of amplifier to get the 100W of power.
3) A PC and Labview to tie it all together.
I cant help you with the hardware (maybe some one else here can.) If you
need help
with the Labview portion, this is a good place to ask.
Hope this was in some small way helpful.
Kevin Kent

Similar Messages

  • LabVIEW and custom controls

    Hi all,
    I'm an italian PhD student working with Polytechnic of Bari.
    I'd like to submit to this forum a question about LabVIEW and custom controls in form of ActiveX components.
    Some introductory word about my problem will be helpful: in the past few months here at Polytechnic of Bari my colleagues and I developed some useful ActiveX controls with custom functionalities to use in LabVIEW.
    For example we developed a "contraves", an
    "endless knob" and a useful counter with a scale very similar to that of standard gazometers or energy counters.
    However we are experimenting some difficulties with the integration of these controls in LabVIEW; the controls are correctly imported in our VIs with using the ActiveX Container but...
    in the block diagram they appear as ActiveX references and we must use property nodes to access even their most useful property (value).
    Please, do someone of you know if (and eventually how) is it possible to integrate more deeply the controls in the LabVIEW environment?
    Do someone of you know if it exist an hypothetical "Custom Controls SDK" for LabVIEW more advanced of that integrated in LabVIEW 6.0?
    Any hint would be appreciated.

    > Please can you specify in more detail how I can obtain the ActiveX
    > refnum input automatically wired to a constant (I have observed this
    > behaviour with the 3D-Graph control as you suggested). I haven't found
    > any reference in the LabVIEW on-line help about the topic "merge-VI".
    >
    The first step is to build a VI that contains the panel and diagram
    elements that you want to drop all at once. In your case, drop the
    ActiveX control at a good size, and on the diagram, you can wire it to
    helper VIs, property nodes, etc. The VI's icon and name are important
    as that is what will show up in the palette menu representing your
    control. The connector pane isn't used.
    With the VI constructed, you want to add it to the palette menu. In the
    Tools menu, Advanced, there is an item about half way down for
    Customizing the Palette menu. You will need to build a New Palette Set
    using the ring and name it whatever you like. You can now add your VI
    wherever you like. Most likely you will want to right click to add a
    new subMenu in the Controls palette and place it in a new .mnu file. In
    the subMenu, right click to add Control. The file dialog will open
    allowing you to select a control file to add to the palette. By
    changing the filter from .ctl to all, you can insert VIs into the
    controls palette.
    All VIs inserted into the Control palette will act as merge VIs, meaning
    that when dragged to a panel, the panel elements will be merged into the
    destination panel, and the diagram elements will be merged into the
    destination diagram.
    For others that want to have these on the Functions palette rather than
    the Controls, you add the VI to the Functions palette, then popup on the
    palette item and check the menu item to make it a Merge VI.
    When it comes time to distribute, you will want to give out your .mnu
    file and the VI of course. Easiest is to place both of them in
    user.lib. If you want the items to go at the top level of the palettes,
    then you can add them to add-ons inside of vi.lib, but this causes more
    upgrade challenges.
    Hope that helps.
    Greg McKaskle

  • Beginner question regarding 32-bit mixing and mixdown workflow

    Hello
    I have a beginner question regarding 32-bit mixing and mixdown.
    If I edit some 16Bit, 44.1kHz Stereo WAV Files and put them into multi-track view to do crossfades, how should I do the mixdown?
    Audition shows me in multi-track view, that it is doing 32-Bit mixing.
    Can I just mixdown to 16Bit, 44.1kHz Stereo without any dithering (as the files are 16Bit, 44.1kHz Stereo to begin with), or will I lose quality that way?
    I will be performing a normalization to 96% to the mixdown and then split to tracks in Audition, as in the end I want to to have an audio CD.
    I guess I could mixdown to 32Bit, then normalize and in the end save back to 16Bit, 44.1kHz Stereo WAV (with dithering, I suppose?), but I want to avoid any unnecessary converting steps.
    Greetings

    Any time you do any processing on a 16bit file in 16 bit only it will degrade the audio slightly due to rounding of the calculations. Working in 32 bit floating point (Audition's default) takes account of all bits generated due to processing.
    So it is always best to work in 32 bit, even if your originals were 16, all the way through until the last stage of saving the files for CD burning. Any losses due to conversion will be insignificant against those due to working 16 bit.

  • Questions on Saving and mining data with Labview and DIAdem

    Hi,
    I am sampling two signals at 200k sampling rate. I am trying to save the data on harddisk and analyzing the data using DIAdem.
    If I save the data using mesurement file format .tdm and .lvm, the file size will be about 4 Gigabytes for only 10 mins' acquisition. It is very slow to process it.
    I used the software Clampex and pCLAMP(Axon Instruments) before. At the same 200k sampling rate and also acquiring two signals, these programs save the data as .atf format and the size is only 400 MB for 10 mins' acquisition.
    I wonder if there is also a good way to handle this situation using Labview and DIAdem, and how to do it?

    Hi, Jonathan:
    I tried the TDM binary file format. The file size is 800 MB per min acquisition and it will take a long time to just open these files. For my application, I have to take data for several hours. I am looking for a way to reduce the size of files.
    Is there any other type of file format that can reduce the size significantly and can be handled easily?

  • Merging labview and mathscript array question tia sal2

    Greetings,
    I really like mathscript functionality it’s just taking a
    while to learn how to merge labview and mathscript together.  I can get an XY graph out which plots the
    correct wave function (thanks to the forums help) the problem is going from 1-D
    Array of cluster 2 elements to 1-D of waveform Double so I can get a sound representation
    of the waveform. 
    Anyone have any suggestions.
    Tia sal2
    Attachments:
    mathscript formula to sound with waveform graph.vi ‏248 KB

    Sorry not sure what happened internet gremlins maybe...here's the vi and graphic image
    Second try Hopefully this will work
    Attachments:
    mathscript formula waveform graph.vi ‏1 KB
    image_of_mathscript_to_labview.jpg ‏108 KB

  • Questions about LabVIEW

    Dear support team,
    Could someone please respond to Eric's questions? Eric works at Bonneville Power Administration and they are gearing up for another major LabVIEW and DSC project. He may have LabVIEW 7 beta. I am on my way out of town for two weeks and won't have access to phone or email. Thanks!!
    Tricia Lee
    DSM Inland Northwest
    [email protected]
    "Pierce, Eric - TNCB-TPP-2"
    05/01/2003 08:05 AM
    To: [email protected]
    cc:
    Subject: The list of question i never sent
    Sorry for the long delay, things have been very busy here with this new project. I have questions / concerns about the DSC, OPC and LOGOS; first here is our architecture.
    We will have two servers running windows 2000 server. Both servers will be live and need to have all DSC components installed. One server is to act as the primary server and the second is to act as a failover server. We have custom components that manage all of our custom services, if one of the custom services stops responding the failover service will shut down all custom services on the primary server and start all of the custom services on the secondary server, thus, the secondary server becomes our primary live server.
    My understanding is that we will have to load a *.spf file on each HMI that tells the HMI where the server is, at the same time we need to initialize the DSC and OPC with *spf and *.lpd files. Now I might have the file extensions wrong but the idea is the same. We have all of our data in a SQL 2000 database, how do we make sure that all of the configuration files match the SQL 2000 db? We have several GUI's that can edit the SQL 2000 data, how do we update the DSC / OPC configuration files?
    So in short here are my current questions:
    How can we manage failover with two live servers
    How can we ensure that DSC / OPC / Logos configuration files match the data in SQL 2000?
    Does any of this make sense? This labview stuff is very new to me.
    Eric Pierce
    ACS Group, Inc.
    Sr. Software Engineer
    w. (360) 619-6284

    Hi Eric,
    You got it right that once the primary machine goes down, the standby machine detects this, and then loads the appropriate SCF file, etc. Caveat: This would mean however that the hardware in question is accessible by both the machines (it's on an Ethernet, for instance). Any hardware physically connected to the primary machine will obviously be not accessible by the secondary (unless there's a way to have a "Y" connection).
    The logging part you mention is not very clear to me. Several questions come to mind: Isn't DSC the logger, logging to its Citadel database? Or, is SQL 2000 the main database? And DSC is logging to SQL 2000? If yes, how? Or is the SQL 2000 querying Citadel and hence getting the data into itself?
    Regards,
    Khalid

  • Problem with Labview and an ARM Cortex

    Good morning,
    I am currently trying to use Labview with a board from ST Microelectronics (MCBSTM32) with an ARM Cortex Processor.
    I use the SDK and have followed the tutorials.
    But, when I try to launch the program (the simple loop as write in the tutorial n°2 : http://zone.ni.com/devzone/cda/tut/p/id/7029 ), Keil gives me an error via Labview:
    "Argument 'DARMSTM' not permitted for option 'device'."
    It seems that Keil does not allow an ARM Cortex from ST as the device.
    Moreover, after having this problem, I am unable to use a Keil project ,even a project which worked before, without Labview. I need to restart the computer.
    I also try to launch the Keil project generated by Labview, without using Labview, and it works. But as soon as I use Labview, I have the error.
    Did anyone already have this error or know how to solve it
    Thank you for your answer and sorry for my bad english.
    Regards,
    Raphaël VAISSIERE

    Hi Raphi,
    So let me make sure I understand,
    The project created in LabVIEW errors out with the message "Argument 'DARMSTM' not permitted for option 'device" 
    If you open the same project in Keil uVision, it runs fine
    Here are my questions:
    1. So how does the code run when run through Keil? does it deploy and run fine?
    2. Did you follow the porting procedure completely?
    Your target STM32F103RB  is techincally supported by Keil but you need to port the RTX kernel to it. This paragraph explains it:
    To determine if your target already supports the RTX Real-Time Kernel, browse to the \Keil\ARM\Startup directory, then browse to the folder that corresponds to the manufacturer of your ARM microcontroller. If there is an RTX_Conf*.c file for your target, then the RTX Real-Time Kernel has already been ported for your ARM device. If no such file exists, skip to chapter 4 for more information on the RTX Real-Time Kernel and a guide for porting RTX to your ARM microcontroller. 
    You also need to port the Real-Time agent to it.
    I just want to make sure that you have followed the guidelines. If you have and are still having problems, we will continue to explore this.
    Thanks,
    National Instruments
    LabVIEW Embedded Product Support Engineer

  • Problems with Labview and Windows XP with error -214721786​5

    Hello. Im a new beginner in Labview and am having a problem with a error message
    I dont understand. Im running Labview on a pc-logger with windows XP.
    The program is written by a guy on my university, and logs voltage from
    a solar cell panel. He doesnt have the time to check in to it, so Im supposed to
    find out what's wrong.
    The error message is as follows :
    Error -2147217865 occured at Cmd Execute.vi->DB Tools Insert
    Data.vi->Solcelle.vi
    Possible reasons :
    Exception occured in Microsoft OLE DB Provider for ODBC Drivers,
    [MySQL][ODBC 3.51 Driver][mysqld-3.23.38-log]Can't open file :
    'solcelledata.MYD'. (errno: 145). in Rec Create - Command.vi-
    >Cmd Execute.vi->DB Tools Insert Data.vi->Solcelle.vi
    I really hope someone can tell me what's wrong and specifically what to do
    with it, or possibly guide me to someone who could suggest a solution to this
    problem.
    Reguards
    chrisdee

    Hi Chrisdee,
          (.. not to discount what TimoS has suggested, however...) 
    Embedded in the error information you provided:
    ... Can't open file : 'solcelledata.MYD'.
    Sounds like the program needs this file?  Has the program been moved to a new location?
    just a wild guess!
    Cheers
    Message Edited by Dynamik on 10-06-2005 04:53 AM
    When they give imbeciles handicap-parking, I won't have so far to walk!

  • Communicat​ing between LabVIEW and Basic Stamp 2.0

    Hello,
    We are trying to control the movement of a Parallax Boe-Bot (robot with
    Board of Education) via LabView programming.  The BOE operates on
    Basic Stamp programming to control the pulse width sent to two servo
    motors.
    Is there a way to establish communication between LabVIEW and Basic
    Stamp 2.0?   We need to send commands from LabVIEW to the
    Basic Stamp to control the pulse width and therefore the rotation of
    the two servo motors.  We also need to send a signal from an
    accelerometer based on the robot to the LabVIEW ADC. 
    The ultimate goal is to do this with wireless communication, but for
    right now we are simply trying to figure out any way to send the data
    between LabVIEW and the Basic Stamp.
    Any ideas?
    Thank you,
    Adam

    Hi Adam,
    You can communicate with a BASIC Stamp microcontroller over GPIB using VISA.  To see some LabVIEW example code, just search for GPIB in the NI Example Finder (Help>>Find Examples). This will show you the basics of GPIB communication in LabVIEW.  I also found this link by searching the NI discussion forums for "basic stamp"
    http://www.parallax.com/html_pages/downloads/softw​are/software_basic_stamp.asp
    Here you can download an example using LabVIEW 7.0 or higher to communicate with a BASIC Stamp 2.
    I hope this is helpful!  Let us know if you have further questions.
    Megan B.
    National Instruments

  • Large applications - Labview and other programming languages

    Hello Labview Users,
    as the forum saw this very interesting thread about large applications programmed in Labview
    (see: http://sine.ni.com/niforum/niforum?requireLogin=False&forumDU=http://forums.ni.com/ni/board/message?... ) I would like to ask the community about their experiences with Labview applications in combination with other programming languages.
    In advance: I have several years of experience in programming Labview applications starting from quick-and-dirty solutions which had to run within few hours and complex test solutions. I saw Labview growing and becoming better with the released versions and lot of things I missed in former times got implemented in the meantime. Actually I have to develop a complexe ATE solution with numerous equipment to control and numerous data to be captured and archived. Despite the Verison 8 I still feel still some drawbacks of the LV concept which let me hestitate to setup the solution completly in Labview:
    1) It is alway hard to re-use code of complex applications since it is not possible to do some kind of global search an replace of functions
    variables etc. It nearly impossible to re-use approved code structures (e.g a state machine) if the "inner part" is changing more the a little bit.
    2) If the application requires a certain flexibility (e.g. exchangeable test equipment of varying vendors) this is hard to implement since you have to define a lot of parameters through your hierarchy if you dont want use global variables which slow down your application and hide
    the code functionality.
    3) Despite modern PCs the look and feel of LV applications appears somewhat slow compared to other applications. For complex user interfaces the polling methode generates a lot of complex code. (I dont have expierence with the event-structure).
    Now my questions:
    Do you have experience of implemention of complex solutions dividingthe code modules using Labview and other languages? Which other
    languages did you use? Why did you use these languages (speed, flexibility of text based code, available library functions)? Did you found out this to improve your development time and code maintainibility?
    (I concider a solution where I do the single tests with Labview-VIs but delegate all the test sequencing and data collection stuff written in PERL which allows really very compact code)
    I'm curious what your experiences are.
    rainercats

    Given that you're asking these questions in a forum for LabVIEW users the opinions given are going to be somewhat slanted towards the general like of LabVIEW. I've been working with LabVIEW for a long time, ever since 2.something on a Mac. I've written numerous large-scale applications as well as "mundane" instrument drivers. As you've noted you're experienced with LabVIEW, so you know some of its strengths and weaknesses.
    To address your specific questions:
    (1) Yes, that has always been a limitation in LabVIEW, but I don't believe it is an overriding one to choose C over LabVIEW. Putnam provided one workaround for the search and replace of VIs. Once you've programmed in LabVIEW long enough you get used to doing it this way. Is it clumsy? Yes. As for the re-use of code structures, that's not entirely true. You can create a "template" VI (a regular VI, not a .vit) that contains the code you want to re-use and place in your palette with the "Merge VI" option set. That way you can select it from your functions palette, plop it down on your diagram, and you get the "template" VI's diagram placed right into your new VI.
    (2) This is not something that is specific to LabVIEW, as this exists with any programming language. It's not the language that limits you here, it's how you've designed your code. In a language like C++ you would go with classes. You can do the same thing in LabVIEW. IVI is another option (though not preferred by me).
    (3) As Putnam mentioned, you should be using the event structure.
    Other thoughts:
    The biggest strength I see with LabVIEW is that each VI is a miniature program, which allows development and debugging of functions a snap. With a language like C you have to write another program to call that function in order to debug it. The biggest weakness? I would say user interface. Yes, even with the event structure. Don't get me wrong, the event structure has vastly improved the way user interfaces and event handling in general are done with LabVIEW, but it simply doesn't hold muster to a program written in C or VB. ActiveX anyone? LabVIEW still doesn't do ActiveX properly in terms of actually getting the controls to work. Programming ActiveX controls is just plain aggravating what with all the property nodes taking up so much diagram space.
    It certainly makes sense to use the best tools available to you to get the job done. In my recent projects I had to write software to run automated tests on some products my company made. The test modules were written in LabVIEW. The tests executive was a proprietary engine driven by a SQL Server database. I had to write a "wrapper" DLL that interfaced between the LabVIEW code and the executive since the executive hadn't been designed to call LabVIEW DLLs directly. This allowed us to use LabVIEW as the preferred language for developing the test modules and let the guys who were fiddling with the test executive do their bit. Of course, TestStand's premise is basically that.

  • Reference Library for Converting Between LabVIEW and XML Data (GXML)

    Please provide feedback, comments and questions on the Reference Library for Converting Between LabVIEW and XML Data (GXML) in this thread.
    The latest version of the NI GXML Library is availble in VIPM on the NI LabVIEW  Tools Network repository.

    Francesco, Thank you for the feedback.  With this component it was my intention to make a more "terse" version of the LabVIEW Flatten to XML VI that was also supported on RT and that gave the user more flexbility regarding the structure of the parsing type definition. I think you are right that the XML parser is not compliant to section 2.11 of the XML spec.  The parser does specifically looking for a #D#A and this appears to be an oversight on my part.  Please confirm for me, the specifcation is saying that the XML parser should be able to recognize three possibilities as an "end of line" character: #D#A, #D, or #A.  Am I reading this right?There are more efficient (and in some cases much more efficient) ways of sharing data between LabVIEW and LabVIEW: some examples are flattened binary strings and the datalog binary format.  XML is slower than these optons but the upside is that it is human readable.  Furthermore XML is inherently hierarchical which is convenient for complex data structures like clusters of arrays of clusters, etc.  If you don't care about human readability then you are correct XML doesn't make as much sense.I will return to the GXML source code and try to fix this in the near future but I would hope that instead of creating yet another custom VI from scratch that you could reuse what I have provided for you.  I included enough documentation in the source code so that users could make some modificiations themselves. The target application for this reference library was LabVIEW to LabVIEW communication.  As such I documented the schema on the dev zone document from a LabVIEW perspective.  It includes all the supported datatypes and all the supported data structures (cluters, arrays, multidimensional arrays, clusters of multidimensional arrays, etc.)  I do see some value in making a more conventional XML spec but the time investment required didn't really line up with my intended use case. Were there any other downsides to GXML that I have missed?  Best Regards, Jeff TippsSystems Engineer - Sound and VibrationMessage Edited by Jeff T. on 04-21-2010 10:09 AM

  • Please help me with my electrical engineering homework : temperature control and watering system for greenhouse using labview and arduino

    temperature control and watering system for greenhouse using labview and arduino
    spesification :
    1. max temp : 28 celcius (when temperature is more than 28 celcius, fan ON)
    2. min temp : 20 celcius (when temperature is under 20 celcius, heater ON)
    3. watering system : aquaponic (grow plant and fish in separate tank but connect each other). Plant roots help filter water for fish. Fish poop for plants fertilizer. So I need a pump to distribute water.
    Please help me create VI file simulation.. I'm sorry I'm not fluent in English. May God bless you all
    Attachments:
    YOOOSHH.vi ‏88 KB

    Duplicate thread.   Please keep the discussion in that thread where you already have a response. It is also the more appropriate thread for your question.
    Lynn

  • Use NI LabView and LabView RT with third party PXI chassis

    I want to use NI LabView and LabView RT, some PXI chassis of N.I and
    some PXI chassis of Third Party( not N.I ) for diversity, How I will Install
    Pharlap ETS ( LabView RT ) on those third party PXI's.
    More simply
    Third party PXI chassis and controllers support LabView RT or not?
    Note: Mark Andre Hary [email protected] of Sensing Systems LLC answer to this question as:
    "You can't.  I think NI bought Pharlap, and they do not offer board support packages (BSPs) for non-NI PXI controllers.  There is no way to get LabVIEW-RT running on a controller not purchased from NI.  You will have to use a standard RTOS like Linux,VxWorks, or QNX.
    Regards

    As described in the web links below, the only way to run LabVIEW Real-Time for ETS on a third-party system would be to follow the procedure for desktop PC computers. You may find that your third party PXI controller may meet the requirements listed in the below link and you will be able to run Real-Time on the controller. If you decide to do this, you will need to purchase the LabVIEW Real-Time Desktop Deployment license (also linked below). After purchasing the deployment license, you will be sent boot disks that will install the real-time OS onto your system. There is no need to purchase Pharlap ETS from Ardence directly.
    http://digital.ni.com/public.nsf/3efedde4322fef19862567740067f3cc/cb3434da94466dd286256e9e006bb775?OpenDocument
    http://sine.ni.com/nips/cds/view/p/lang/en/nid/13751
    Instead of using LabVIEW Real-Time for ETS, you could use LabVIEW Real-Time for RTX instead. This system allows you to run Windows and the Ardence Real-Time eXtensions (RTX) real-time kernel concurrently on one machine. Currently device support for LabVIEW RT for RTX systems is limited to RIO devices. This is all described in the below link.
    http://zone.ni.com/devzone/conceptd.nsf/webmain/9cc60298315a687e86256e78007b3a96
    I hope this answers your question.
    Regards,
    Kristi H - National Instruments

  • How could I copy an image under labview and to stick it right A side of old

    copying shapes and stick them x times after knowing that one will differemment modify the color of each shape, I have Imaq Vision Builder

    I'm not sure I understand what you are trying to achieve. Will you please clarify the question? Also, what version of LabVIEW and Vision Builder are you using?

  • Beginner Question

    Hello all, hopefully I have posted this to the correct LiveCycle board!
    I'm new to Acrobat and LiveCycle in general. I'm working out of windows LiveCycle Designer 8.x  and my question is as follows....
    I'm working off of the "Quote" Form Template and what I am looking for is a way to make the rows of the Table expand as needed.  I will be converting this Template into a "Purchase Order" form, and there will be many occassions where I will need to increase the table rows - perhaps onto a second or even third page - due to the volume of different items I will be seeking prices on at a given time.
    Long story short I would like to take this "QUOTE" form template and make it such that if needed, I can expand the table size one row at a time, pushing the content beneath (Comments, subtotal, etc) onto the next page....
    Hoping I have explained my query well enough,  Could anyone provide me a solution or point me in the correct direction?
    Many thanks!

    As a followup, so as not to clog the board with numerous threads I have another beginner question
    For the sake of hypotheticals I am creating a dynamic PDF form.  This pdf is a Purchase Order Form, with a table in the Body that will expand/contract as the form user needs.  (EXA-  Default table size is 5 items, however user is creating an order for 12 items, he/she can increase the table rows to desired capactiy.) It will be a very simple layout. All data for the form will come from manual entry, no XML data binding going on.
    My new question is as follows:    Is there a way I can have this user's Adobe reader or Adobe acrobat software remember previous data entries, so that if 4 weeks after ordering widget X he/she needs to place another order for widget X, the software will autofill the table as he/she types into the table cell?
    This would be similar to how Excel will start to autofill if it recognizes a previously entered value from some other cell in the worksheet (workbook?)
    EXA-
    [NAME]   [MODEL]   [QTY]   [PRICE]
    Widget           A            433        $30.00
    Would it be possible for the software to notice that I'm typing the word "widget" somewhere after typing "wid" ???
    Many thanks for any and all followups, I'm a teenager trying to help out in my father's small business. They are in the darkages here and I'm doing my best to self teach.
    Thanks.

Maybe you are looking for