Execution time of a simple vi too long

I'm working with LabVIEW 6.0.2 on a computer (AMD ~700MHz) under Windows 2000. The computer is connected to the instruments (eg Keithley 2400 Sourcemeter) via GPIB (NI PCI-GPIB Card). When trying to read the output of the K2400 with a very simple vi (sending the string READ? to the instrument with GPIBWrite (mode 2) and subsequently reading 100byte with GPIBRead (mode 2) from the instrument, the execution time mostly exceeds 1s (execution highlighting disabled). Sometimes, it can be much faster but this is very irreproducible. I played around with the GIPBRead and Write modes and with the number of bytes to be read from the device as well as with the hardware settings of the Keithley 2400 but nothing seemed to work. The API calls ca
ptured by NI Spy mainly (lines 8 - 160) consist of ThreadIberr() and ibwait(UD0, 0x0000).
As this problem is the main factor limiting our measurement speed, I would be grateful for any help.
Thanks a lot
Bettina Welter

Hello,
Thanks for contacting National Instruments. It seems like the 1 second delay that is occurring is due to the operation being called. ThreadIberr returns the value of iberr, while ibwait simply implements a wait. These two get called repeatedly while the GPIB device waits for the instrument (K2400, in your case) to finish its operation and respond back. It is quite possible that when you query the Keithley to send back 100 bytes of data, it has to gather them from its buffer (if its already been generated). And if there aren't 100 bytes of data in the buffer, the Keithley will keep the NRFD line asserted while it gathers 100 btyes of data. After the data has been gathered, the NRFD line is deasserted, at which point, ThreadIberr will detect the change in th
e ibsta status bit and read the 100 bytes.
So make sure that the 100 bytes of data that you are requesting don't take too long to be gathered, since this is where the main delay lies. Hope this information helps. Please let us know if you have further questions.
A Saha
Applications Engineering
National Instruments
Anu Saha
Academic Product Marketing Engineer
National Instruments

Similar Messages

  • I have the converter to change pdf to word. Works sometime, other time says conversion is taking too long. Can I get what it has translated so I can use the parts?

    I have the converter to change pdf to word. Works sometime, other time says conversion is taking too long. Can I get what it has translated so I can use the parts?

    I would ask in the ExportPDF forum, Adobe ExportPDF (read only) (assuming that's the service to which you have subscribed). This is the Reader one.

  • Time machine backups take way too long

    Greetings,
    Having trouble with Time Machine backups taking way too long. (No, it's not the first backup after some momentous event, as 90% of what I read about slow backups mentions; it's _every_ backup.)
    20" iMac purchased Oct 2008. OS 10.6.8. 500 GB internal hard drive. 3 TB external backup drive. Was tested at the local "Genius" Bar, but they found nothing wrong with it. The best they could do is to recommend a restore from Time Machine, and if that didn't work, start clean and reload the OS and then copy my stuff back. I don't think so. I actually did a full Time Machine restore for a different reason, and it didn't help (except that I gained a few GB of disk space!).
    My Time Machine backups now take about 39 minutes to complete. I have a 500 GB internal drive and this problem happens regardless of which external drive I use. (I have 3: 2 WD hard drives and one [noisy!] G-drive.) My current backup drive is a 3 TB WD drive with 2 TB of free space.
    These are small incremental backups, typically about 35 MB. It doesn't take half an hour to copy 35 MB!
    All phases take a long time, but the most irritating one is where TM appears to get stuck for a long time at the beginning and end of the copying phase. For example, at one point it will say something along the lines of Backing Up 7 KB of 34.4 MB. And it'll be like that for minutes on end. Later it will say something like 35 MB of 35 MB and hang there for minutes on end. During these periods the write I/O rate will go up and down, and CPU usage will often be very high (with backupd using 90 - 100%, in addition to highly elevated System CPU usage). In one particular backup the I/O write rate for one part of it looked like a heartbeat on an EKG! I should have taken a screenshot.
    At the beginning during the Calculating Changes phase, backupd is using between 90 and 100% CPU.
    There is also high CPU usage during the next phase, Preparing. And the number of items being prepared is in the thousands.
    It always goes in two rounds because it finds that the root directory (/) has been modified.
    I've Googled and searched Mac forums and found nothing useful. I started with a new .plist file and nothing's changed.
    Can anyone help? Thanks!
    I wrote the above Jun 28. Yeah, finally getting around to posting it. Since then it's gotten worse. I now do a manual backup once a day and it takes almost an hour. I tried re-indexing Spotlight. No good.
    Can anyone help? Again, this is not a "first backup." It is _every_ backup.
    Thanks!
    AEF

    Thanks for your efforts. Here are my answers, comments, and further info:
    I believe I have repaired permissions on my drive as part of my troubleshooting. And I'm sure I checked SMART status, which came up with a thumbs up, so to speak. I'm trying again. I don't see any useful messages in any logs on the console. I'm looking at "All Messages" and there is so far nothing useful. Oh, I think you must mean repair as in "Repair Disk," not "Repair Disk Permissions." I believe the "genius" at the Apple store did that or something just as good. The machine checked out fine according to them.
    Checked SMART status just now: passed. Ran Quick Drive Test: also passed.
    Oh, I installed "Time Machine Buddy". Occasionally, and only occasionally, I get a "deep traversal needed message" like this (I copied this from .Backup.log):
    Running preflight for "Macintosh HD" (mount: '/' fsUUID: BF606E9A-5FD0-3B92-8D76-33DC63E7B2B1 eventDBUUID: 30DD929D-8982-46C2-BAE6-F0BD1E73916A)
            Scanning nodes needing deep traversal
            Node requires deep traversal: /Users/alanfeldman/Pictures/iPhoto Library/Database/apdb reason:contains changes|must scan subdirs|fsevent|missed reservation|
            Calculating size of changes
            Should copy 2919 items (27.8 MB) representing 7116 blocks of size 4096. 546805181 blocks available.
    Preflight complete for "Macintosh HD" (mount: '/' fsUUID: BF606E9A-5FD0-3B92-8D76-33DC63E7B2B1 eventDBUUID: 30DD929D-8982-46C2-BAE6-F0BD1E73916A)
    Time elapsed: 1 minute, 39.000 seconds
    But the vast majority of the time I do NOT get this message. But if there's something wrong with this apdb file, what can I do to fix it?
    Note that there is usually, if not always, a lot of CPU activity or I/O (disk) activity, but not a lot getting done.

  • Time machine back up taking too long

    Why is time machine taking so long to back up?  I just downladed the update and it's still dragging

    You are entitled to 90 day telephone support from Apple. Try contacting them.
    Apple Support Contact
    Apple Support contact - Telephone

  • Time to open Web forms too long for the first connexion !!

    Hi
    When we first open a web form in the day, the time to open it, is greater than 10 minuts ( the CPU time used to retrieve the webform is measured at about 20 seconds)
    When we do open the webform a second time it takes about 20 seconds to open it!!
    Have someone meet this case?
    What do we have to change in order to enhance the openning duration of the webforms?
    Thanks
    environment:
    Serveur IBM, OS Windows 2003 server
    Hyperion planning 11.1.1
    Edited by: RYAG on 26 avr. 2010 14:56

    Does it take 10 minutes to open on the server, if it takes 10 minutes on the server and only 20 seconds of that is CPU processing then are you saying for over 9 minutes it is rendering the form in your browser, you can check the browser process on the machine.
    If you do want to check or increase the maximum heap size then have a read of :- Re: Heap sizes with Planning11.1.1.3 & Weblogic help
    I would also check how much memory the JVM is using when opening the form, you can check using task manager on the server or better still use something like process explorer.
    Cheers
    John
    http://john-goodwin.blogspot.com/

  • Transaction execution time and block size

    Hi,
    I have Oracle Database 11g R2 64 bit database on Oracle Linux 5.6. My system has ONE hard drive.
    Recently I experimented with 8.5 GB database in TPC-E test. I was watching transaction time for 2K,4K,8K Oracle block size. Each time I started new test on different block size, I would created new database from scratch to avoid messing something up (each time SGA and PGA parameters ware identical).
    In all experiments a gave to my own tablespace (NEWTS) different configuration because of oracle block-datafile size limits :
    2K oracle block database had 3 datafiles, each 7GB.
    4K oracle block database had 2 datafiles, each 10GB.
    8K oracle block database had 1 datafile of 20GB.
    Now best transaction (tranasaction execution) time was on 8K block, little longer tranasaction time had 4K block, but 2K oracle block had definitly worst transaction time.
    I identified SQL query(when using 2K and 4K block) that was creating hot segments on E_TRANSACTION table, that is largest table in database (2.9GB), and was slowly executed (number of executions was low compared to 8K numbers).
    Now here is my question. Is it possible that multiple datafiles are reasone for this low transaction times. I have AWR reports from that period, but as someone who is still learning things about DBA, I would like to asq, how could I identify this multi-datafile problem (if that is THE problem), by looking inside AWR statistics.
    THX to all.

    >
    It's always interesting to see the results of serious attempts to quantify the effects of variation in block sizes, but it's hard to do proper tests and eliminate side effects.
    I have Oracle Database 11g R2 64 bit database on Oracle Linux 5.6. My system has ONE hard drive.A single drive does make it a little too easy for apparently random variation in performance.
    Recently I experimented with 8.5 GB database in TPC-E test. I was watching transaction time for 2K,4K,8K Oracle block size. Each time I started new test on different block size, I would created new database from scratch to avoid messing something up Did you do anything to ensure that the physical location of the data files was a very close match across databases - inner tracks vs. outer tracks could make a difference.
    (each time SGA and PGA parameters ware identical).Can you give us the list of parameters you set ? As you change the block size, identical parameters DON'T necessarily result in the same configuration. Typically a large change in response time turns out to be due to changes in execution plan, and this can often be associated with different configuration. Did you also check that the system statistics were appropriately matched (which doesn't mean identical cross all databases).
    In all experiments a gave to my own tablespace (NEWTS) different configuration because of oracle block-datafile size limits :
    2K oracle block database had 3 datafiles, each 7GB.
    4K oracle block database had 2 datafiles, each 10GB.
    8K oracle block database had 1 datafile of 20GB.If you use bigfile tablespaces I think you can get 8TB in a single file for a tablespace.
    Now best transaction (tranasaction execution) time was on 8K block, little longer tranasaction time had 4K block, but 2K oracle block had definitly worst transaction time.We need some values here, not just "best/worst" - it doesn't even begin to get interesting unless you have at least a 5% variation - and then it has to be consistent and reproducible.
    I identified SQL query(when using 2K and 4K block) that was creating hot segments on E_TRANSACTION table, that is largest table in database (2.9GB), and was slowly executed (number of executions was low compared to 8K numbers).Query, or DML ? What do you mean by "hot" ? Is E_TRANSACTION a partitioned table - if not then it consists of one segment, so did you mean to say "blocks" rather than segments ? If blocks, which class of blocks ?
    Now here is my question. Is it possible that multiple datafiles are reasone for this low transaction times. I have AWR reports from that period, but as someone who is still learning things about DBA, I would like to asq, how could I identify this multi-datafile problem (if that is THE problem), by looking inside AWR statistics.On a single disc drive I could probably set something up that ensured you got different performance because of different numbers of files per tablespace. As SB has pointed out there are some aspects of extent allocation that could have an effect - roughly speaking, extents for a single object go round-robin on the files so if you have small extent sizes for a large object then a tablescan is more likely to result in larger (slower) head movements if the tablespace is made from multiple files.
    If the results are reproducible, then enable extended tracking (dbms_monitor, with waits) and show us what the tkprof summaries for the slow transactions look like. That may give us some clues.
    Regards
    Jonathan Lewis

  • SQL statements in Orchestration take too long

    Hello all,
    I have an orchestration that inside has a Expression Widget.
    In this Expression widget I call several methods from an assembly that I wrote in VstudioC#.
    This assembly get some values from the files beeing processed and does some checking (empty values, accepted values, etc..) and does some select/insert queries into a separate database  that I use for tracking/analysis.
    These sql statement from some time now are taking just too long, taking each file around 5 minutes inside the Expression and some of the sql statement taking as long as 1 minute.
    I've cleaned up the tables from my separate database, as they had more that 2 million records.
    The queries are simple, without joins , etc...
    Any ideas on hot to make this work faster?

    Why are you doing SQL Operations in an Expression Shape?
    The standard BizTalk Pattern would be by the WCF SQL Adapter.

  • How to get the execution time of a Discoverer Report from qpp_stats table

    Hello
    by reading some threads on this forum I became aware of the information stored in eul5_qpp_stats table. I would like to know if I can use this table to determine the execution time of a worksheet. In particular it looks like the field qs_act_elap_time stores the actual elapsed time of each execution of specific worksheet: am I correct? If so, how is this value computed? What's the unit of measure? I assume it's seconds, but then I've seen that sometimes I get numbers with decimals.
    For example I ran a worksheet and it took more than an hour to run, and the value I get in the qs_act_elap_time column is 2218.313.
    Assuming the unit of measure was seconds than it would mean approx 37 mins. Is that the actual execution time of the query on the database? I guess the actual execution time on my Discoverer client was longer since some calculations were performed at the client level and not on the database.
    I would really appreciate if you could shed some light on this topic.
    Thanks and regards
    Giovanni

    Thanks a lot Rod for your prompt reply.
    I agree with you about the accuracy of the data. Are you aware of any other way to track the execution times of Discoverer reports?
    Thanks
    Giovanni

  • Backups taking too long

    My Time Capsule backups are taking too long. Time Machine says it is indexing. However, I have had my Time Machine configured since I installed Lion last year and it has been making automatic backups every hour since. Why is it 'indexing' now, and why is it taking so long?
    Thank you

    Visit pondini.org for all things Time Machine.

  • Portal report execution time

    Hi experts,            
    Can any one tell me how to check the report execution time in portal.Its taking a long time to execute
    the report in portal.so i would like to evaluate the execution time in portal itself.
    I know how to check it at bi side but here iam intrested to know at portal itself and it varies
    at bi side RSRT and at portal level. At RSRT and in Analyzer its taking expected time only but in
    portal its taking long time.
    Regards,
    Shais

    Thanks Alot For your reply.
    Its working. when i run the statistics i got the below details .
    Step runtime: 7532 ms   -
    > is this the total time taken to execute the report?
    Remote runtime :6845 ms
    Local runtime :687 ms
    Handle types olap frontend events:
      - BRFC: 78 ms
      - DP: 237 ms
      - OLAP: 6443 ms
      - DFLT: 95 ms
      - DIAL: 0 ms
      - APPL: 391 ms
      - F4: 4 ms
      - CMD: 110 ms
      - ITEM: 109 ms
    what is meant by remote run time, local run time,BRFC .......... Can you explain me in details.
    Waiting for your reply,
    Regards,
    Shais.

  • OPM process execution process parameters takes too long time to complete

    PROCESS_PARAMETERS are inserted every 15 min. using gme_api_pub packages. some times it takes too long time to complete the batch ,ie completion of request. it takes about 5-6 hrs long time ,in other time s it takes only 15-20 mins.This happens at regular interval...if anybody can guide me I will be thankful to him/her..
    thanks in advance.
    regds,
    Shailesh

    Generally the slowest part of the process is in the extraction itself...
    Check in your source system and see how long the processes are taking, if there are delays, locks or dumps in the database... If your source is R/3 or ECC transactions like SM37, SM21, ST22 can help monitor this activity...
    Consider running less processes in parallel if you have too many and see some delays in jobs... Also indexing some of the tables in the source system to expedite the extraction, make sure there are no heavy processes or interfaces running in the source system at the same time you're trying to load... Check with your Basis guys for activity peaks and plan accordingly...
    In BW also check in your SM21 for database errors or delays...
    Just some ideas...

  • Query taking too long a time

    Hi friends,
    the view I have created on a table is taking toooo long a time to gimme the results
    when am trying to select * from table_name
    Can some one suggest me a solution pls
    CREATE OR REPLACE VIEW XXKDD_LATEST_SAL
    (RN, MONTH, YEAR, EMPLOYEE_NUMBER, POSITION,
    PAYROLL_NAME, DEPT, STATUS, TERMINATION_DATE, FULL_NAME,
    TOP3, BASIC_SALARY)
    AS
    select "RN","MONTH","YEAR","EMPLOYEE_NUMBER","POSITION","PAYROLL_NAME","DEPT","STATUS","TERMINATION_DATE","FULL_NAME","TOP3","BASIC_SALARY" from (
    SELECT ROW_NUMBER() OVER (PARTITION BY employee_number ORDER BY employee_number) rn, tp.*
    FROM (SELECT MONTH, YEAR, employee_number, position, payroll_name, dept, status, termination_date, full_name,
    ROW_NUMBER () OVER (PARTITION BY employee_number, basic_salary ORDER BY YEAR , MONTH) top3,
    DECODE (basic_salary,
    100000, 4500,
    24000, 1921,
    basic_salary
    ) basic_salary
    FROM kdd_pay_hr_sal_vw
    order by employee_number,year desc) tp
    WHERE top3 <= 1
    select * from XXKDD_LATEST_SAL

    Read these informative threads:
    When your query takes too long ...
    HOW TO: Post a SQL statement tuning request - template posting
    And edit your post, add relevant details like database version, execution plan etc., it is all listed in the above links.
    And use the {noformat}{noformat}-tags, to keep your examples formatted and indented and readable.                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                                       

  • Discoverer report taking too long time to open.

    HI,
    Discovere reports are taking too long time to open. Please help to resolve this.
    Regards,
    Bhatia

    What is the Dicoverer and the Application release?
    Please refer to the following links (For both Discoverer 4i and 10g). Please note that some Discoverer 4i notes also apply to Discoverer 10g.
    Note: 362851.1 - Guidelines to setup the JVM in Apps Ebusiness Suite 11i and R12
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=362851.1
    Note: 68100.1 - Discoverer Performance When Running On Oracle Applications
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=68100.1
    Note: 465234.1 - Recommended Client Java Plug-in (JVM/JRE) For Discoverer Plus 10g (10.1.2)
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=465234.1
    Note: 329674.1 - Slow Performance When Opening Plus Workbooks from Oracle 11.5.10 Applications Home Page
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=329674.1
    Note: 190326.1 - Ideas for Improving Discoverer 4i Performance in an Applications 11i Environment
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=190326.1
    Note: 331435.1 - Slow Perfomance Using Disco 4.1 Admin/Desktop in Oracle Applications Mode EUL
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=331435.1
    Note: 217669.1 - Refreshing Folders and opening workbooks is slow in Apps 11i environment
    https://metalink.oracle.com/metalink/plsql/ml2_documents.showDocument?p_database_id=NOT&p_id=217669.1

  • Sometimes my computer takes too long to connect to new website. I am running a pretty powerful work program at same time, what is the best solution? Upgrading speed from cable network, is it a hard drive issue? do I need to "clean out" the computer?

    Many times my computer takes too long to connect to new website. I have wireless internet (time capsule) and I am running a pretty powerful real time financial work program at same time, what is the best solution? Upgrading speed from cable network? is it a hard drive issue? do I only need to "clean out" the computer? Or all of the above...not to computer saavy.  It is a Macbook Pro  osx 10.6.8 (late 2010).

    Almost certainly none of the above!  Try each of the following in this order:
    Select 'Reset Safari' from the Safari menu.
    Close down Safari;  move <home>/Library/Caches/com.apple.Safari/Cache.db to the trash; restart Safari.
    Change the DNS servers in your network settings to use the OpenDNS servers: 208.67.222.222 and 208.67.220.220
    Turn off DNS pre-fetching by entering the following command in Terminal and restarting Safari:
              defaults write com.apple.safari WebKitDNSPrefetchingEnabled -boolean false

  • HT201250 my time capsule is taking too much time indexing backup and then taking longer time to back up ( 207 days ) or longer !!! what shall i do ?

    my time capsule is taking too much time indexing backup and then taking longer time to back up ( 207 days ) or longer !!! what shall i do ?

    Try 10.7.5 supplemental update.
    This update seems to have solved this problem for many.
    Best.

Maybe you are looking for

  • Configuring external Seagate drive so that I can use Time Machine on two separate machines

    I am using both an iMac and a Laptop (Macbook Pro) and wish to use Time Machine to back up both machines. However, in my ignorance it seems not to recognise which machine it backs up and thus overwrites itself from the 'other' machine. Can I do this

  • How do I add a scroll bar to an array?

    I want to take a 2D array and change the row its is in with a scrollbar so that in a another array index I can pull out individual number from each row.  So I have a CSV file [time,1,2,3,4,5] and I want the user to able to vary the time with a scroll

  • How to use a class in Flash

    I'm trying to use a class (LoadSwf.as) in Flash. The code is listed at the bottom. Here's my code in Flash: import LoadSwf var myLoader:LoadSwf=new LoadSwf() addChild(myLoader) The file being imported (Login1.swf) contains a very simple ComboBox. Whe

  • Apple push notification not working in production(Ad Hoc Certificate)

    We are totally stucked, please help. I and my team made a iPhone application, and this is the first time we try on iOS. Everything is fine, until we submitted our app and became available on app-store -- the push notification service is not working.

  • Service contract list for plant

    Hi How to query from a table for the Service contracts created for a particular plant. I tried EKKO table , but could not get for a particular plant. Can anyone guide me how to get the list. Thanks in advance Gururaj