How to control HP 33120A and output a DC signal?

Hellow!
I'd like to control HP 33120A function generator,right now i have use the HP 33120A-WFM Config EX.VI and i can output Sin,Squart...wave,but when i try to output the DC signal,the amplitude control seems not work,what shall i do to output a DC signal and control its amplitude,Thanks!

I haven't worked with this generator or its driver, but sometimes the DC level is controlled by changing the offset instead of the amplitude. You might want to check this out.
Hope that this helps,
Bob
Bob Young - Test Engineer - Lapsed Certified LabVIEW Developer
DISTek Integration, Inc. - NI Alliance Member
mailto:[email protected]

Similar Messages

  • How to control (the input and output) EXE file after I call it using exec?

    Hi,
    I knew that I can use runtime.exec() to call one EXE file, and this works. But this EXE has two characteristics:
    1. After this exe starts, it asks user to input number such as 1 or 2 onto computer screen, then press return. Then the exe will start the calculation.
    2. after it starts calculation, it prints 3 columns of numbers onto the screen.
    My two questions are:
    1. How to use java to input the number such as 1 or 2 automatically? this EXE can not work like this in DOS command line:
    C:> file.exe parameter
    The parameter is the number such as 1 or 2 that I wanna input.
    2. how to redirect the 3 columns of numbers from computer screen to txt file?
    My colleague can solve these two questions using Mathematica. So I know that definitely there is at least one solution for it. I just can not do it using Java. This wierd exe file bothered me a lot and I really wish that I can get help from someone in java community.
    Thank you!
    Tony

    When you call Runtime.exec, you get a Process object. (I presume something similar happens when you use ProcessBuilder.) Process has methods with names getOutput, getInput, and getError. These correspond to the standard input, standard output, and standard error streams of the spawned process.
    You can read and write to the process on the streams corresponding to input and output that the process writes to the console.
    [add]
    In fact, you should be grabbing and reading the output/error streams anyway, because of the points raised by the Traps article. Google "Java Runtime exec traps" and you'll probably get a link to this JavaWorld article, which describes common Runtime.exec problems and how to solve them.
    Edited by: paulcw on Jun 15, 2010 4:09 PM

  • How to change the input and output schema in BPEL process

    hi',
    Please tell me how to change the input and output schema in BPEL process after the process is made.
    thanks
    Yatan

    If your intention is just changing the content you are passing to bpel/returning from bpel
    Here is another way
    just update your default created xsd files with new elements, update wsdl elements in message definition and chnage bpel code to reflect new elements in activities
    Regards,
    Praveen

  • How to view the input and output layouts created in the planning folder?

    Hi all,
    How to view the input and output layouts created in the planning folder in the bw 3.5?
    Thanks
    Pooja

    Hi,
    You can refer to this link. How to create planning folder and executing the planning folder.
    http://help.sap.com/saphelp_nw04/Helpdata/EN/5d/7c4b52691011d4b2f00050dadfb23f/frameset.htm
    Hope this helps

  • How to set input delay and output delay when source Synchronous

    ClkIN is the board clock which is connected to the FPGA. Clkif is the generated clock from ClkIN. The Device's clk come from Clkif. So, how to set input delay and output delay in this scene(within my understand, this is Source Synchronous)?
    The example in many document, the input delay and output delay setting all refer to board clock(within my understand, this is System Synchronous). In that scene, the input delay max = TDelay_max + Tco_max; input delay min = Tdelay_min + Tco_min; the output delay max = Tdelay_max + Tsu; output delay min = Tdelay_min - Th.
    So, I want to know how to set input/output delay in the Source Synchronous.
    In system synchronous, I set input/output delay such as:
    create_clock -period 20.000 -name ClkIN -waveform {0.000 10.000} [get_ports ClkIN]
    create_generated_clock -name Clkif -source [get_pins cfg_if/clk_tmp_reg/C] -divide_by 2 [get_pins cfg_if/clk_tmp_reg/Q]
    create_clock -period 40.000 -name VIRTUAL_clkif //make virtual clock
    set_input_delay -clock [get_clocks VIRTUAL_clkif] -min 0.530 [get_ports DIN]
    set_input_delay -clock [get_clocks VIRTUAL_clkif] -max 7.700 [get_ports DIN]
    set_output_delay -clock [get_clocks VIRTUAL_clkif] -min -0.030 [get_ports DOUT]
    set_output_delay -clock [get_clocks VIRTUAL_clkif] -max 1.800 [get_ports DOUT]
    *******************************************************************************************

    So, first. Architecturally, the clock that you forward to your external device should not come directly from the clock tree, but should be output via an ODDR with its D1 input tied to logic 1 and the D2 tied to logic 0. This guarantees minimal skew between the output data and the forwarded clock.
    ODDR #(
    .DDR_CLK_EDGE("OPPOSITE_EDGE"), // "OPPOSITE_EDGE" or "SAME_EDGE"
    .INIT(1'b0), // Initial value of Q: 1'b0 or 1'b1
    .SRTYPE("SYNC") // Set/Reset type: "SYNC" or "ASYNC"
    ) ODDR_inst (
    .Q (Clkif_ff), // 1-bit DDR output
    .C (ClkIN_BUFG), // 1-bit clock input
    .CE (1'b1), // 1-bit clock enable input
    .D1 (1'b1), // 1-bit data input (positive edge)
    .D2 (1'b0), // 1-bit data input (negative edge)
    .R (rst), // 1-bit reset
    .S (1'b0) // 1-bit set
    OBUF OBUF_inst (.I (Clkif_ff), .O (Clkif_out));
    This generates an output clock that is the same frequency as your input clock. This is consistent with your drawing, but inconsistent with your constraints - is the forwarded clock a 50MHz clock or a 25MHz clock?
    I will assume your ClkIN goes to a BUFG and generates ClkIN_BUFG.  Your first constraint generates a 50MHz clock on the ClkIN port which will propagate through the BUFG to (among other places) this ODDR.
    create_clock -period 20.000 -name ClkIN -waveform {0.000 10.000} [get_ports ClkIN]
    Assuming your forwarded clock is supposed to be 50MHz, then your 2nd command is close to correct
    create_generated_clock -name Clkif -source [get_pins cfg_if/ODDR_inst/C] -combinational  [get_pins cfg_if/ODDR_inst/Q]
    With this done, you have successfully described the forwarded clock from your design. This is the clock that goes to your device, and hence should be the clock which is used to specify your input and output constraints.
    set_input_delay -clock [get_clocks Clkif] -min 0.530 [get_ports DIN]
    set_input_delay -clock [get_clocks Clkif] -max 7.700 [get_ports DIN]
    set_output_delay -clock [get_clocks Clkif] -min -0.030 [get_ports DOUT]
    set_output_delay -clock [get_clocks Clkif] -max 1.800 [get_ports DOUT]
    If you want to get fancier, you could try adding a set_clock_latency to the forwarded clock to account for the board propagation of the clock
    set_clock_latency -source TDtrace2 [get_clocks Clkif]
    (But I haven't experimented with clock latency on a generated clock and I don't know for a fact that it works).
    Avrum

  • How to set the input and output of a subVI

    I am trying to configure a subVI by following Chap. 7 of "LabVIEW fundamentals". But I stuck at "Building the connector pane".
    My subvi is quite simple, just for testing purpose. I plan to do a summation and display the result. So I placed one numeric Add icon on the block diagram and a numeric Indicator on the front panel. I built a connector pane with three terminals, two for input and one for output. 
    But, how could I associate the terminals on connector pane with inputs and outputs of terminals on the block diagram? I wanna build a subVI, whose input are two numbers and display the summation on the subVI.
    Forrest Sheng Bao, Ph.D.
    Assistant Professor, Dept. of Electrical & Computer Engineering
    University of Akron, Akron, OH, USA
    https://sites.google.com/site/forrestbao/

    Hi Forrest Bao,
    use the right mouse key as often as you can in LabVIEW. It´s very helpful. If you right click on a link in the connector pane, then you can select to disconnect selected or all.
    Mike
    Message Edited by MikeS81 on 08-28-2008 01:33 PM

  • How to know the printprogram and output type for perticular standard script

    how to know the print program and output type for perticular standard script and wt the proceure we will fallow to modifie the standard script..if possible u have any sample code plz share with me.....

    Hi Phani,
    To know the print program and output type for a standard script we have a table called TNAPR. So just you go to SE11 and provide the table name as TNAPR and choose display option.Then go with display entries option and in selection screen you have to provide the form name for whic you want to know the print program and output type.
    To modify a standard script first of all you have to copy the standard form for which you need modifications into Zform. Once you done this you can modify the layout set as per your requirement.Then what ever the modifications you have done in layout set for that you need to populate the data.For this you also need to do some modifications to the print program.so you have to options like directly you can modify the print program as per your requirement and other one is you can create a subroutine and do coding in that subroutine. Then you need call this subroutine in the Script Form.like this you can do modifications to the standard script........................... Reward me if its helpful..........
    Thanks & Regards.
    Laxman.P
    B'lore.

  • How to Control the Valves and Pump with Labview?

    I have started a new project where I have to control the valves operation through Labview 2009,using PXI 1031, NI 5105(Digitizer) and NI 6251(M-series DAQ).
    The valves I need to control using the labview are:
    a) Pressure Reducing Valve (open & close)
    b) Remote control Valve(On & Off) with in-built actuator
    c) Flow control valve(open, semi open, partially open, semi close, partially close, fully close), 1%, 2%, 3%..etc
    d)saftey valve (opening closing depending on increase in pressure from certian limit)
    I hav eto control
    a) Hydraulic pump operation with in built gear-box
    Please let me know, how should l develop the VI for all? How to get started?, any helpful links? If there is any example code?
    Its little urgent !

    If you are not familiar with LV, I would suggest you go through the tutorials. Then take the leap and figure out how to use Action Engines and State Machines (search any of the terms I used that don't understand, there are a zillion threads on them).
    Then...
    Sit down and write stand-alone testers for each of the widgets you are going to control monitor etc.
      These testers will give you an idea if they really work the way you think, their character, and will serve as a "Sanity Check" when you think you are going crazy  durring development.
    THen integrate the tester so they can all work at the same time. This will become a utility screen that will serve you well when maintaining the critter latter in its life.
    Only then...
    write out a complete set of state diagrams that describe what you plan to do (don't cheat! ) and walk through them in your head until you are cetain it will all work blah blah blah.
    This step is optional but highly recomened...
    Compose a detailed set of documemtns describing what you have in mind and post them here (include the images, they will get many mor readers) and let people  pick aprat your design.
    When you have a final design, use everything you learned along the way to be a kick-ass developer.
    Along the way post distinct well defined question on this site and as long as it looks like you are trying someone will probably reply and get you going again.
    Ben
    Ben Rayner
    I am currently active on.. MainStream Preppers
    Rayner's Ridge is under construction

  • How to save terminal command and output history to pdf on exit?

    When I exit a terminal session (terminal.app), I would like to automatically save the commands and output from the session to a PDF (or text file) as a reference. 
    I can do this manually in Terminal by selecting Shell --> Print... --> PDF --> Save as PDF...
    Unfortunately, I'm too forgetful to remember to do this each time before I exit. 
    Is there an alias I could set or some other method to automatically save the contents of my session to a PDF when I exit?
    Thank you in advance,
    jjw

    Frank,
    Your response was helpful, but I realized after looking at my .bash_history file that only the commands entered are listed, not the output.  I would like my commands to be printed out, but also the result from the commands (in this simple example, the directory listing):
    myuserid[~/galaxy-dist/tools]$ cd ncbi_blast_plus
    -------------------------------------------------------------------------------- ----------------------------------------------------- 09:48:37
    myuserid[~/galaxy-dist/tools/ncbi_blast_plus]$ ls
    blastxml_to_tabular.py   hide_stderr.py           ncbi_blastp_wrapper.xml  ncbi_tblastn_wrapper.xml
    blastxml_to_tabular.xml  ncbi_blastn_wrapper.xml  ncbi_blastx_wrapper.xml  ncbi_tblastx_wrapper.xml
    -------------------------------------------------------------------------------- ----------------------------------------------------- 09:48:38
    myuserid[~/galaxy-dist/tools/ncbi_blast_plus]$
    I'm thinking that there is an automator script or an alias that I could use so that when I type, "exit", the commands and output for the session would be sent to a PDF and then the session would close.
    I know it seems like an odd thing to want, but sometimes I'll forget the sequence of commands that gave me the correct output (plus, my command prompt has a timestamp on it, too).
    Thanks, though, for the .bash_history information.
    jjw

  • How to control Color Mode and Resolution for brochure creation?

    Hello guys,
    I am trying to find some settings within Pages, where I could set CMYK color mode and 300 dpi resolution for my pdf file.
    I am creating brochure from scratch, using blank document, because non of brochure templates work for what I need. But I can't find any controls for color mode and resolution set up.
    Is there any solution for this?

    There has been a lot of discussion in the Pages forum about resolution and some on CMYK.
    Much of the discussion has been incorrect.
    For the colour handling there is the colour palette
    The Apple Colour Palette is the user interface for colour managing vector objects created in Apple Pages such as rules and panels and for colour managing one class of vector objects not created in Apple Pages, namely typographic objects.
    Imported objects maintain their colourant model and colour space, if there is one. If there is none, a colour space is assigned and the assigned colour space is saved out into the PDF, opposite e.g. Adobe Photoshop which will colour manage the object in the open editing session, but will then save out the object to disk without the assigned ICC profile if 'Colour Management: Off' is configured.
    Therefore, if the imported object is data space CMYK and the colour space device CMYK, the system will assign the Generic CMYK Profile which is equivalent to ANSI CGATS TR001 colourimetry (US SWOP); if the imported object is data space RGB and the colour space is device RGB, the system will assign the Generic RGB Profile which is equivalent to sRGB colourimetry in OS X 10.5 and higher.
    In general, always capture and then convert for correction in one of two data spaces with three channels, either ICC colour managed RGB or ICC CIELa*b D50. Then convert to CMYK1, CMYK2, CMYK3, CMYKn from the ICC colour space chosen for correction and archiving. This colour space should be large enough to hold any colours that the output colour spaces can hold.
    The Apple ColorSync Utility has a gamut comparison dialogue which is 3D (point, press, and turn the 3D gamut comparison). Use this facility to pick an ICC colour space of reasonable size and shape for correction and conversion into final output. Don't pick sRGB as a quick look will show that it can't hold as input what ISO reference offset can hold as output for pure cyan, let alone hold what a big inkjet can hold in greens, blues, reds and yellows.
    The ICC architecture lets the ICC profile for the printed output match out of gamut colours to in gamut colours using the Perceptual rendering intent, as intended in the original architecture from 1995. Adobe internally introduced an inhouse conversion for compatibility with PostScript and as default for PDF, Relative Colorimetric with Adobe Black Point Compension. Relative Colorimetric is in fact a proofing transform, without gamut mapping. Black Point Compensation is a simple linear scaling of the lightness to compensate for the fact that the shadows will be clipped in a Relative Colorimetric conversion.
    The other problem is transparency, introduced in the QuickDraw Graphics Extension for Mac OS 7.5. Transparency is not supported in any version of Adobe PostScript (level 1, level 2 or level 3). Transparency is also not supported in any version of Adobe PDF prior to version 1.4 and higher. Therefore, unless the RIP Raster Image Processor is able to accept PDF 1.4 and higher, transparency has to be rendered / flattened which is done in the PDF/X-3 filter of the Apple ColorSync Utility.
    The idea in the PDF/X-3 filter (for Adobe PDF 1.3) is that the everyday enduser, after placing ICC colour managed RGB or CIE La*b D50 objects, and after choosing an ICC colour space for objects created in Pages, chooses the ICC colour space for the output condition. The filter accepts both ICC profiles of type MNTR Monitor with data space RGB, ICC profiles of type PRTR Printer with data space RGB, and ICC profiles of type PRTR Printer with data sapce CMYK.
    The colourant conversion (RGB to CMYK) and the gamut mapping (out of gamut colours in input colour spaces mapped to in gamut colours in output colour space) is then calculated as per the ICC profile for the output condition. An output condition is the combination of factors that affect how colour is formed: the colour lightness and colour tint of the paper, the colours of the colourants, and the tone value increase. Think of gamut mapping in a good ICC profile as Format 1 kerning in a TrueType font (where your font has state-based kerning for up to 8 glyphs simultaneously).
    Below a link to a discussion of how to change the default PDF/X-3 filter the Apple ColorSync Utility. "Wegen diverser Fehler im Mac OS X empfiehlt es sich, wo immer möglich, die PDFs ohne Transparenzreduktion (also auf normalem Weg) über den Befehl Als PDF sichern zu erstellen und das Flattening der Druckerei zu überlassen," that is, pick the plain save as PDF path and leave the transparency flattening to the prepress technician.
    Henrik
    Reference:
    http://www.icalamus.net/tutorials/howtocreatepdfx.php

  • How to synchronize analog input and output from two different USB daq boards

    Hi all,
    I have two very differnt USB boards the NI USB 6008, which I am using to acquire the data (Analog Input) and a NI USB 9263, it is an Analog Output only board that I am using to deliver a signal (in this case a square pulse). The reason why I am not using the 6008 Analog Ouputs is because I need to deliver negative voltages and need the full +/-10V range.
    Looking at similar posts I am pretty sure that I can't use an external trigger or a shared clock, I also tried to use the synchronization of timed structures but no cigar.
    I am including a quick vi that I whipped out showing how the signal jitters due to the lack of synchronization. The AO from the 9263 connects to the AI in the 6008 in this example.
    Solved!
    Go to Solution.
    Attachments:
    Test Pulse.vi ‏117 KB

    I talked to a specialist in the phone and tols me that it is not possible.

  • I recently installed your new version of firefox but i still cannot see the security settingsfirefox is experiancing a problem will have to close down or how to control viruses etc and as your answer

    all i want to know is how can i check for viruses etc using firefox and where can i find the security settings or are you doing that automatically,it has also crashed when i opened a vcl player

    To downgrade to Firefox 3.6 first uninstall Firefox 4, but do not select the option to "Remove my Firefox personal data". If you select that option it will delete your bookmarks, passwords and other user data.
    You can then install the latest version of Firefox 3.6 available from http://www.mozilla.com/en-US/firefox/all-older.html - it will automatically use your current bookmarks, passwords etc.
    To avoid possible problems with downgrading, I recommend going to your profile folder and deleting the following files if they exist - extensions.cache, extensions.rdf, extensions.ini, extensions.sqlite and localstore.rdf. Deleting these files will force Firefox to rebuild the list of installed extensions, checking their compatibility, and reset toolbar customizations.
    For details of how to find your profile folder see https://support.mozilla.com/kb/Profiles

  • How to display error messages and output from Matlab (which Matlab would typically send to its command window but no longer does when called by Labview) into Labview or allow it to be dumped into Matlab Command Window?

    Using Labview 6i and Matlab 6.1. I want to be able to see Matlab warnings and error messages either in the Matlab Command Window or in Labview itself. Currently Matlab is called by Labview (which is working). However I would like to debug and/or modify my Matlab script file to better understand how the two programs are interfacing. It is difficult since no data or messages can be displayed currently to the Matlab command window. I would like to change that if it is possible - Labview is suppressing that from happening. If not possible to send these
    messages to Matlab Command Window can I make it at least possible to see Matlab's actual warnings and/or error messages in Labview?

    I don't think you can debug your Matlab script from labVIEW. The following webpage talks about this:
    http://digital.ni.com/public.nsf/3efedde4322fef198​62567740067f3cc/19106e318c476e608625670b005bd288?O​penDocument
    My suggestion would be to write a script in Matlab and thoroughly test it before calling the script from a Matlab script node in LabVIEW.
    Chris_Mitchell
    Product Development Engineer
    Certified LabVIEW Architect

  • How to control the size and position of the region on a page?

    I am trying to create welcome page for my client. I need to create a banner and place the login region in the cente of the page with a fixed width and height. How can I do this.
    Thanks.
    Satya

    Hi,
    Edit your region and go to region header of your region
    and type this
    region header
    <table width ="100px" height = "100px">
    <tr>
    <td>
    region footer
    </td>
    </tr>
    </table>
    you can customize your region position in this way...just use table tag attributes,,i hav egiven example of only height and width but you can use all the table attributes to customize your region like background color,back ground image,alignment etc..
    hope this helps you,
    Regards,
    Jitendra

  • How to display aux bus and Output 1-2 as tracks in arrange?

    Sorry for the silly question but I would like to display the auxiliary busses and Out 1-2 as "tracks" in the Arrange view. They show in the Mixer as channel strips and, of course, they're in the Arrange Inspector but I would like to be able to see all my Busses and final Out at the same time. I've gone through the manual to no avail so it must be something really simple I'm missing. Can someone help me out?
    Thanks.

    Simplest method is to click on the 'Mixer' tab in your main arrange page. That should bring up the track mixer, then in the mixer tab's "Options" menu you'll find "create arrange tracks for selected channel strips".
    Another method, which I'm actually kinda unhappy about, is that if you create any automation for a bus or aux chananel, it automatically creates an arrange track for the bus/aux. Doesn't seem like such a bad idea in itself, except that it totally messes with your track mixer display order. For example, I have all my subgroup bus strips together in my track mixer. But if I create some automation for one of'em, like the vocal bus for example, suddenly it's track strip jumps to somewhere else in the mixer. The only way I can keep bus tracks together is to create 'dummy' automation for all of'em - which is what I have to do.
    I wish you could manually move the strips in the track mixer around to your preferred order, and choose to 'de-link' them from the arrange view. I've submitted this already to Apple's Black Hole. I mean, the 'Feedback' page...

Maybe you are looking for