Transaction Rate Per Second

How can I determine the transaction rate per second for my database? I found the following script:
SELECT SUM(s.value/
(86400*(SYSDATE - TO_DATE(i.VALUE,'J')))) "tps"
FROM V$SYSSTAT s, V$INSTANCE i
WHERE s.NAME in ('user commits','transaction rollbacks')
AND i.KEY = 'STARTUP TIME - JULIAN' ---
But - I don't know where the "86400" value came from? Does anyone have any insight? - or a valid script to determine the tps (transaction rate per second)? Thanks

Dianne,
There are 86400 seconds in a day. The query you are using divides all commits and rollback transactions by the time the system has been up and running in seconds. Hence, transactions per second...... tps.....
Hope this is an accurate response.
Adam
<BLOCKQUOTE><font size="1" face="Verdana, Arial">quote:</font><HR>Originally posted by Dianne Jones ([email protected]):
How can I determine the transaction rate per second for my database? I found the following script:
SELECT SUM(s.value/
(86400*(SYSDATE - TO_DATE(i.VALUE,'J')))) "tps"
FROM V$SYSSTAT s, V$INSTANCE i
WHERE s.NAME in ('user commits','transaction rollbacks')
AND i.KEY = 'STARTUP TIME - JULIAN' ---
But - I don't know where the "86400" value came from? Does anyone have any insight? - or a valid script to determine the tps (transaction rate per second)? Thanks <HR></BLOCKQUOTE>
null

Similar Messages

  • Counting pulses and measuring flow rate per second

    Hi everyone
    I am using a compactField Point (cFP-2220) controller to measure the flow rate. The flow sensor that i have, gives the output in terms of the pulses. I need to count the number of pulses for a particular interval and divide it by a K-factor( constant value) to get the flow rate. If i select the interval to be of a second, i would want to count number of pulses per second and divide it by K-factor to get flow rate ( per second). 
    How do i count pulses per second and consequently measure flow rate using this count and K-factor.
    Any kind of help would be highly appreciated.
    Thanks in advance.

    You may be able to use the Edit Tempo function to work this out. With you waveform in Edit view Right Click on the time scale at bottom of waveform and select Bars and Beats display. Select a number of beats by highlighting with the cursor. Then again on the Time bar select Edit Tempo. In that dialogue you can try clicking on Extract to find the Tempo ie. beats per minute. Also you can try adjusting the tempo by entering avalue manually to see how close that gets to your waveform.
    Haven't used this function of Audition much so you may have to refer to the help file to find out more about it.

  • Oracle 10g AQ - Required Insert Rates  of atleast 8000+ inserts per second

    Hey Guys,
    I am setting up a AQ in oracle 10g and I require atleast 8000+ insert rates per second, is it possible if so How do i implement it.
    Any help would be helpful.
    Thanks
    Stan

    Can you Give me some examples :
    I am trying to use AQ Messaging : How do I use perl scripts for the AQ messaging.
    In any case if you can give me some examples; detailed on how you have coded
    to attain the insert rates of 8000 messages per second
    I would appreciate that
    Thanks
    stan kumar

  • Calculate transaction per second

    Dear Friends ,
    I have run the Oracle 10g with AIX unix server . I want to know "How many Transaction is occured per minute or second" in PEAK hour into my production server . How can I find it ? Give me some suggestions plz ... ..

    An Oracle "transaction" isn't the same as Business "transaction". If you are getting "transactions per second" to report to management (even if it is IT management), you better be sure you identify which transactions you are counting.
    If I run a PLSQL loop that does
    loop 1 to 1000
      insert into table ...
      commit ;
    end loopI get 1,000 "transactions" and 1000 INSERT "executions" being reported by Oracle.
    On the other hand if I run
      loop 1 to 1000
        insert into table ...
    end loop
    commit;I get only 1 "transaction" and 1000 INSERT "executions" being reported by Oracle.
    Next, if I run
      insert into table ... select ... 1000 rows ..
    commit;I get 1 "transaction" and 1 INSERT "execution".
    Any database application will have a mix of "transactions" of different sizes being executed by different users / application server processes / clients.
    If you take the aggregate "transactions" count you are going to be sorely disappointed and disappointing your managers !

  • Monitor Transactions Per second

    Some times I see my SQL Server is running less than 2000 Transactions per second (from the "Transactions /sec" Perfmon counter) and some times I see there are 17000 Transactions per second.
    If I want to find out which application is executing these transactions or which query is executing these transactions, is there a way to find this out?

    If its a prod and a busy server, running the profiler *for few hours* wont be a good idea. You may opt to run the sever side traces with only specific events.
    http://msdn.microsoft.com/en-us/library/cc293613.aspx
    If you are using SQL server 2012 you can use extended events
    http://msdn.microsoft.com/en-IN/library/bb630282.aspx
    Satheesh
    My Blog |
    How to ask questions in technical forum

  • My Apple Mail is showing 100 million incoming messages and climbing at a rate of about 20k per second

    My Apple Mail is showing 100 million incoming messages and climbing at a rate of about 20k per second. Where are all these going and how do I get this to stop?
    thanks

    Whoa, was your email exposed on a public forum/TV show?
    First I would sign out of Mail right away.
    Second I would go to your email's website and sign into there. Examples... icloud.com, hotmail.com, yahoo.com and whatever you use. Just to see if it's happening there. If it is happening there, then I strongly suggest changing your email address and whatever website  you used it on for login information.
    If it's not happening there, then hopefully when you sign back into Mail it won't happen again. If it does happen again, let us know. I wouldn't be able to help but maybe someone else could.
    KOT

  • Control the transaction per second ( TPS ) to down stream system through Oracle SOA

    Hi,
      As part of our business flow,  BPEL calling down stream system via OSB.
    However the down stream system can accept only 5 request per second.
    1. Tried implementing DB poller to invoke down stream system each second with 5 records.
    The flow is as follows .
    DB --> Poller --> ParentBPEL --> Down Stream System
    In this case Poller can able to poll 5 records from DB and push it to "DownStreamSystem" via "ParentBPEL" process.
    But here the contraint is , the "DownStreamSystem" strickly can able to process only 5 requests per second.
    If the "DownStreamSystem" is processed 4 request and processing 1 request  at first second, then in the next second the fusion should push only 4 request . Since the number records in processing state should be 5 at a time.
    Please help me out to solve this scenario or suggest me if there is any other alternative to achieve the solution.
    Thanks in Advance ...

    maybe throttling in the osb is of any help ?
    Throttling
    since you really have hard specs on the "currently processed messages" you could add some coherence cache and use that as sort of lookup to see what messages are getting processed at which time. (it's just some guessing, since i don't have any experience with the situation you're describing)

  • Weird number of disk-transfers per second when in SYNC mode

    Hi, I've managed to configure BDB in synchronous mode (i.e., each put is persisted on disk when committing). However, now I'm doing 2000 puts per seconds, each with payload from 10 to 250 kilobytes, yet I'm getting (from iostat) that each disk-transfer is only about 6 kilobytes (because of 23 megabytes written on disk per second, divided by 3750 transfers per second). How is that even possible? Is there a way of telling BDB to minimize the number of disk-transfers per second in SYNC mode? It seems that BDB is breaking each put's payload into smaller pieces, and only then saving to disk in a bunch of disk-transfers.

    Hi,
    JE does not split up a single write into multiple writes -- and certainly doesn't do an fsync for each one.
    JE may do multiple writes (but not fsyncs) for a single, multi-operation txn if the write buffer fills.  And it will do multiple writes for a single operation, if the record is larger than the write buffer.  However, it doesn't sounds like this (overflowing of the write buffer) is what you're experiencing.  In any case, you can configure the size of the JE write buffer with EnvironmentConfig.LOG_TOTAL_BUFFER_BYTES, LOG_NUM_BUFFERS, and LOG_BUFFER_SIZE.
    Another thing is that JE will group fsyncs (this is called "group commit") when multiple threads are committing concurrently with SYNC durability.  In this case you'll see a smaller number of physical writes than the number of commits.
    I asked a colleague who has more experience with iostat than I do about this, and he gave me the following information:
    We would expect there to be one sync per put on average, assuming the application is doing serial writes and there are no group commits to further obfuscate the issue. Given the high sync write rate, the writes are presumably to an SSD, or spinning rust with a large non-volatile disk write cache
    I'm not sure what he means by disk transactions in iostats. Perhaps he means the number of disk transfer requests issued to the device listed as tps (transfers per sec) in iostat output.
    If he is using ext3 and he does not have the file system mounted with noatime, he may be observing write requests to update the file system "atime" metadata with each request. So for 2K sync puts/sec he would see roughly 4K (2k put + 2k atime update) write requests/sec and his average write payload would be ~12KB/transfer (the atime write payload is negligible), which would be consistent with the application's put behavior. This is all a guess.
    I hope this helps.
    --mark

  • Downloads per Second slower than normal

    Hi, i've moved from profile ADSL MAX to ADSL 2+ however my downloads per second are average around 2-400kbs a sec, my download speed is just over 7mb and my downloads have been around 850kbs a sec.
    Any reason to why this is?
    I don't do much heavy downloads, most of my connection is used for either gaming or downloading some the app or playstation and its really slow at downloading anything.
    If you want to say thanks for a helpful answer, please click on the Ratings star on the left-hand side If the the reply answers your question then please mark as ’Mark as Accepted Solution

    sorry my computer crashed here it is
    FAQ
    Test1 comprises of two tests
    1. Best Effort Test:  -provides background information.
    Download  Speed
    4339 Kbps
    0 Kbps
    7150 Kbps
    Max Achievable Speed
     Download speedachieved during the test was - 4339 Kbps
     For your connection, the acceptable range of speeds is 2000-7150 Kbps.
     Additional Information:
     Your DSL Connection Rate :7192 Kbps(DOWN-STREAM), 1060 Kbps(UP-STREAM)
     IP Profile for your line is - 6345 Kbps
    2. Upstream Test:  -provides background information.
    Upload Speed
    821 Kbps
    0 Kbps
    1060 Kbps
    Max Achievable Speed
    >Upload speed achieved during the test was - 821 Kbps
     Additional Information:
     Upstream Rate IP profile on your line is - 1060 Kbps
    We were unable to identify any performance problem with your service at this time.
    It is possible that any problem you are currently, or had previously experienced may have been caused by traffic congestion on the Internet or by the server you were accessing responding slowly.
    If you continue to encounter a problem with a specific server, please contact the administrator of that server in the first instance.
    If you want to say thanks for a helpful answer, please click on the Ratings star on the left-hand side If the the reply answers your question then please mark as ’Mark as Accepted Solution

  • How many of these objects should I be able to insert per second?

    I'm inserting these objects using default (not POF) serialization with putAll(myMap). I receive about 4000 new quotes per second to put in the cache. I try coalescing them to various degrees but my other apps are still slowing down when these inserts are taking place. The applications are listening to the cache where these inserts are going using CQCs. The apps may also be doing get()s on the cache. What is the ideal size for the putAll? If I chop up myMap into batches of 100 or 200 objects then it increases the responsiveness of other apps but slows down the overall time to complete the putAll. Maybe I need a different cache topology? Currently I have 3 storage enabled cluster nodes and 3 proxy nodes. The quotes go to a distributed-scheme cache. I have tried both having the quote inserting app use Extend and becoming a TCMP cluster member. Similar issues either way.
    Thanks,
    Andrew
    import java.io.Serializable;
    public class Quote implements Serializable {
        public char type;
        public String symbol;
        public char exch;
        public float bid = 0;
        public float ask = 0;
        public int bidSize = 0;
        public int askSize = 0;
        public int hour = 0;
        public int minute = 0;
        public int second = 0;
        public float last = 0;
        public long volume = 0;
        public char fastMarket; //askSource for NBBO
        public long sequence = 0;
        public int lastTradeSize = 0;
        public String toString() {
            return "type='" + type + "'\tsymbol='" + symbol + "'\texch='" + exch + "'\tbid=" +
                    bid + "\task=" + ask +
                    "\tsize=" + bidSize + "x" + askSize + "\tlast=" + lastTradeSize + " @ " + last +
                    "\tvolume=" + volume + "\t" +
                    hour + ":" + (minute<10?"0":"") + minute + ":" + (second<10?"0":"") + second + "\tsequence=" + sequence;
        public boolean equals(Object object) {
            if (this == object) {
                return true;
            if ( !(object instanceof Quote) ) {
                return false;
            final Quote other = (Quote)object;
            if (!(symbol == null ? other.symbol == null : symbol.equals(other.symbol))) {
                return false;
            if (exch != other.exch) {
                return false;
            return true;
        public int hashCode() {
            final int PRIME = 37;
            int result = 1;
            result = PRIME * result + ((symbol == null) ? 0 : symbol.hashCode());
            result = PRIME * result + (int)exch;
            return result;
        public Object clone() throws CloneNotSupportedException {
            Quote q = new Quote();
            q.type=this.type;
            q.symbol=this.symbol;
            q.exch=this.exch;
            q.bid=this.bid;
            q.ask = this.ask;
            q.bidSize = this.bidSize;
            q.askSize = this.askSize;
            q.hour = this.hour;
            q.minute = this.minute;
            q.second = this.second;
            q.last = this.last;
            q.volume = this.volume;
            q.fastMarket = this.fastMarket;
            q.sequence = this.sequence;
            q.lastTradeSize = this.lastTradeSize;
            return q;
    }

    Well, firstly, I surprised you are using "float" objects in a financial object, but that's a different debate... :)
    Second, why aren't you using pof? Much more compact from my testing; better performance too.
    I've inserted similar objects (but with BigDecimal for the numeric types) and seen insert rates in the 30-40,000 / second (single machine, one node). Obviously you take a whack when you start the second node (backup's being maintained, plus that node is probably on a separate server, so you are introducing network latency.) Still, I would have thought 10-20,000/second would be easily doable.
    What are the thread counts on the service's you are using?; I've found this to be quite a choke point on high-throughput caches. What stats are you getting back from JMX for the Coherence components?; what stats from the server (CPU, Memory, swap, etc)?; What spec of machines are you using? Which JVM are you using? How is the JVM configured? What's are the GC times looking like? Are you CQC queries using indexes? Are your get()'s using indexes, or just using keys? Have you instrumented your own code to get some stats from it? Are you doing excessive logging? So many variables here... Very difficult to say what the problem is with so little info./insight into your system.
    Also, maybe look at using a multi-threaded "feeder" client program for your trades. That's what I do (as well as upping the thread-count on the cache service thread) and it seems to run fine (with smaller batch sizes per thread, say 50.) We "push" as well as fully "process" trades (into Positions) at a rate of about 7-10,000 / sec on a 4 server set-up (two cache storage nodes / server; two proxies / server.) Machines are dual socket, quad-core 3GHz Xeons. The clients use CQC and get()'s, similar to your set-up.
    Steve

  • 10,000 Recorc Per Second (In EJB 3.0)

    hi all,
    i have some mission critical tasks into my project, is it possible to persist 10 000 record per seconds,
    1. AS - JBoss Application Server 4.0.4GA
    2. Database - Oracle 10G 10.2.0.1
    3.EJB - 3.0 Framework
    4.OS - SunOS 5.10
    4.Server - Memory: 16G phys mem, 31G swap, 16 CPU,
    i know that i need performace
    here is my configurations about performance
    1. JVM Config Into JBoss
    JAVA_OPTS="-server -Xmx3168m -Xms2144m -Xmn1g -Xss256k -d64 -XX:PermSize=128m -XX:MaxPermSize=256m
       -Dsun.rmi.dgc.client.gcInterval=3600000 -Dsun.rmi.dgc.server.gcInterval=3600000
        -XX:ParallelGCThreads=20 -XX:+UseConcMarkSweepGC -XX:+UseParNewGC
        -XX:SurvivorRatio=8 -XX:TargetSurvivorRatio=90 -XX:MaxTenuringThreshold=31 -XX:+AggressiveOpts
        -verbose:gc -XX:+PrintGCTimeStamps -XX:+PrintGCDetails -XX:+PrintTenuringDistribution2. also i configure my database.xml file
    <?xml version="1.0" encoding="UTF-8"?>
    <datasources>
      <xa-datasource>
        <jndi-name>XAOracleDS</jndi-name>
        <track-connection-by-tx/>
        <isSameRM-override-value>false</isSameRM-override-value>
        <xa-datasource-class>oracle.jdbc.xa.client.OracleXADataSource</xa-datasource-class>
        <xa-datasource-property name="URL">jdbc:oracle:thin:@192.168.9.136:1521:STR</xa-datasource-property>
        <xa-datasource-property name="User">SRVPROV</xa-datasource-property>
        <xa-datasource-property name="Password">SRVPROV</xa-datasource-property>
        <exception-sorter-class-name>org.jboss.resource.adapter.jdbc.vendor.OracleExceptionSorter</exception-sorter-class-name>
        <min-pool-size>50</min-pool-size>
        <max-pool-size>200</max-pool-size>    
        <metadata>
             <type-mapping>Oracle9i</type-mapping>
          </metadata>
      </xa-datasource>
      <mbean code="org.jboss.resource.adapter.jdbc.vendor.OracleXAExceptionFormatter"
             name="jboss.jca:service=OracleXAExceptionFormatter">
        <depends optional-attribute-name="TransactionManagerService">jboss:service=TransactionManager</depends>
      </mbean>
    </datasources>3. Also i have one simple Stlateless Session Bean
    @Stateless
    @Remote(UsageFasade.class)
    public class UsageFasadeBean implements UsageFasade {
         @PersistenceContext(unitName = "CustomerCareOracle")
         private EntityManager oracleManager;
         @TransactionAttribute(TransactionAttributeType.REQUIRED)
         public long createUsage(UsageObject usageObject, UserContext context)
                   throws UserManagerException, CCareException {
              try {
                   oracleManager
                             .createNativeQuery("INSERT INTO USAGE "
                                       + " (ID, SESSION_ID, SUBSCRIBER_ID, RECDATE, STARTDATE, APPLIEDVERSION_ID, CHARGINGPROFILE_ID, TOTALTIME, TOTALUNITS, IDENTIFIERTYPE_ID, IDENTIFIER, PARTNO, CALLTYPE_ID, USAGETYPE, APARTY, BPARTY, CPARTY, IMEI, SPECIFICCALLTYPE, APN, SOURCELOCATION, SMSCADDRESS, MSC_ID, ENDREASON, USAGEORIGIN, BILL_ID, CONTRACT_ID) "
                                       + " VALUES(SEQ_USAGE_ID.NEXTVAL, NULL, NULL, SYSDATE, SYSDATE, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL) ");                              
                   return 1;
              } catch (Exception e) {
    }3. and into client side i have 200 Threads, each of them tried to call this method 50 times
    my result is that i can persist 10000 record in 20 seconds, without hibernate, with hibernate i got worst result :(,
    also i hear that it is good idea to use JDBC 3.0 driver for performance,
    i download newest oracle jdbc jar file from oracle site
    http://www.oracle.com/technology/software/tech/java/sqlj_jdbc/htdocs/jdbc_10201.html
    is this jar file JDBC 3.0 driver ?
    is there any hibernate performance configuration?
    is it any more performance tuning into JBoss or EJB with entity beans?
    can anybody help me ? or is there any doc which can help me ?
    Regards,
    Paata,
    Message was edited by:
    paata
    Message was edited by:
    paata

    What makes you think that your database, just the database (with the box that it is on) can handle that rate?
    What makes you think that your network can handle that?
    While this is going on is this the ONLY traffic that will be on the network?

  • Newbie trying to understand the frame/fields per second concept on video

    Do video camera shutter speeds (1/60sec) reflect an interlaced field or a full frame comprised of two interlaced fields per part of a second.
    I suspect that a shutter setting of 1/60sec means it's not a full frame of video but is just an odd or even lined interlace field. Meaning that if I want to shoot 30fps I need to keep my shutter setting on 1/60sec.
    But if that's really the case, then what am I exactly shooting per second if actual NTSC frame rate is 29.97fps? I mean, the first 2p frames can easily be divided into two fields each, but what about the remaining .97frame. How can you divide that into two fields of interlaced video lines?
    Forgive my ignorance but books have a bad habit of not answering back when you don't understand something they say.
    iMac Intel Duo-Core; Intel Mac mini single-core   Mac OS X (10.4.6)  
    iMac Intel Duo-Core; Intel Mac mini single-core   Mac OS X (10.4.6)  

    The shutter speed is not relevant. I can be either field based or progressive. The frame rate is not dependent on the shutter speed.

  • Final Cut Pro X Image Sequence Export missing frames per second option?

    I am using the trial version of Final Cut Pro X.  I can export an image sequence but it will only allow me to do so at 30 frames per second -every single frame! 
    So a 10 min movie takes half an hour to export 20,000+ frames that no one on earth has time to look thru.
    In my old Final Cut Express I was able to choose 1 or 2 frames per second, which was just right.
    What am I missing?
    Do I need to buy Quicktime Pro or Compressor to allow me to export image sequences without exporting every frame?
    Why would Apple even have an image sequence export if it only allows you to export every frame, or it that only in the trial version?
    Thank you

    I found a free option for you, if you don't already have Compressor 4.
    In FCPX create your movie. Then SHARE as a Master File.
    Get MPEG Streamclip, which is a free app available here.
    http://www.squared5.com/svideo/mpeg-streamclip-mac.html
    Drag your Master File into MPEG Streamclip.
    In MPEG Streamclip, select FILE/EXPORT TO OTHER FORMATS.
    In FORMAT: choose IMAGE SEQUENCE
    In OPTIONS, choose your frame rate.

  • 1,000,000 updates per second?

    How could you configure a coherence cluster to handle processing a million stock quotes per sec? The datafeed could be configured as a single app spewing out all 1,000,000/sec or it could be many apps producing proportionately fewer ticks/sec but in any case it's going to total a million/sec. Fractions of the feed spread among multiple physical servers sounds smartest. The quote Map.Entry would probably have a Key of String (or char[] if that's more efficient - i know the max length). The Value would be a price and a size so maybe just those two elements byte[]{Float,Integer} or a java object with Float and Integer member variables. I'd want to trigger actions based on market conditions when the planets align just right so I'm not simply ignoring these values or pub/sub'ing them out to client apps, I'm evaluating many of them simultaneously and using them as event triggers. Is something like that remotely possible? On how much hardware?
    Thanks,
    Andrew

    Andrew,
    Using partitioning, Coherence can handle 1 million updates per second, but the big question is how many updates per second do you need on the hottest instrument at the hottest time?
    The other question is related to "the planets lining up", because that may imply a global view of the market, which becomes more difficult in a partitioned system.
    To provide a high rate of change to data in a partitioned system, the data providers (those with a large amount of data or a high rate of change) should be in the cluster (not coming in over *Extend) to eliminate one hop. To avoid blocking on the tick update from the data provider, it should locally enqueue the update. The queue servicer (a separate thread) should either coalesce whatever ticks are in the queue into a single putAll(), or if every tick needs to be recorded (i.e. all three ticks in the queue like "change to 3.5", "change to 3.55", "change to 3.6" have to be published, instead of just the latest "change to 3.6") then it would batch up everything in the queue until it hits an item that it already has in its batch, and then do a putAll().
    The use of that async publishing mode is what allows for the much higher throughput, particularly when a data provider is producing a huge number of ticks in a given period of time. You can make it even smoother (e.g. avoid outliers caused by some servers being slower) by having more local queues+services (partitioned by Coherence partition, or at the extreme by instrument). You can determine the Coherence partition using the KeyPartitioningStrategy returned from the PartitionedService for the ticks cache.
    Peace,
    Cameron Purdy | Oracle Coherence
    http://coherence.oracle.com/

  • CO Pulse Frequency doesn't actually generate 1 Pulse per Second?

    Hello all,
    I have a VI layed out in the attachment below.  I seem to have a lack of understanding on how to program this VI here.  I just don't understand what is possibly going wrong.
    The VI is very basic.  The frequency has been set to 1, and the units are Hertz.  To me, this means that the program should send one pulse to my linear actuator ONCE per second.  I have a simple pulse counter set up in the VI as well to count how many pulses are actually being sent (using the DAQ assistant).  Why is it that when I run the program, I get around 300 pulses per second?  Raising the value makes it goes slightly faster, but lowering the value doesn't really make it go any slower.  There seems to be no real correlation between the input frequency and the actual number of pulses that are sent.
    I just simply want a program that I can input "1 pulse per second" or however many pulses I want per second and have the card send ONE pulse per second (or however many is input).  Where do I start?  I have a whole program written out and ready to go, but this basic concept here completely eludes me. 
    Thanks,
    James
    Attached:  1) Picture of concept that I'm completely baffled about  2) VI of my program which said concept is being used in
    Solved!
    Go to Solution.
    Attachments:
    Voltage vs Distance SM v1.1.vi ‏475 KB
    COPulseFreq.jpg ‏75 KB

    In the simple image, you are running a loop as fast as possible (there's no timing mechanism).  Inside that loop, you configure the pulse task, start it, then immediately stop and clear it.  You need to create the channel and configure the timing outside the while loop, before it starts, and you need to clear the task outside the loop as well, after the while loop terminates.  Depending on what you want to do, you may be able to move the task start outside the loop as well, or just let the task auto-start.
    You'll need to restructure your VI a bit.  I can't tell if you want to clear the task after each step, or just change the frequency.  If it's just the frequency, you can use DAQmx Write to change it; if you need to start and stop the task, you'll need some logic to do that once each time you want to restart (you may get an error if you start a task that's already running).  There's no need for "Is Task Done?" since you're not using the output for anything.
    EDIT: Also, it is always a good idea to put a wait inside loops that execute indefinitely.  Otherwise they will spin as fast as possible, consuming all available processor time and preventing other code from running.  If you configure your counter task properly the loop timing won't affect the pulse rate (because that's done in hardware) but there's no need to run the loop that fast.

Maybe you are looking for

  • Crystal Report XI: Can you write your query in SQL instead of using the GUI

    Hello   In crystal report version XI, can you write your query in sqlplus and then use the crystal report designer to build your report instead of using the GUI . I would like to be able to go database and show sql query and open that query and make

  • Can't print to a Canon MF 4690 wirelessly through Airport Extreme

    I cannot seem to figure out why this isn't working. I just replaced a Brother printer (7 years old) that worked great wirelessly through my Airport Extreme with a Canon imageCLASS MF 4690. I found, downloaded and installed the drivers for it (10.6),

  • Trying to send email with contents in the body of the message

    Hi, Has anyone been successful in modifying the RW_EMAIL_NOTIFICATION script in such a way that allows the contents of the email "body" to be sent as a parameter from a calling script? Thanks. Regards, David Carr

  • Issues in generating xml from Report Builder as per XSD provided

    Hi, We are provided with XSD's for each report. Now we need to generate XML's from report that are as per the XSD's provided to us. But the problem come's when we have to define hirearchies. Like for example 1. Our Report is based on the following qu

  • Too many points in a path

    Sorry if this has been posted asked a million times, i couldn't find anything in a search. Received any .ai file with meeellions of points. Anyway to clean it up or get rid of a few. Choking the printer. Thanks in advance for any help.