Count the Pass/ Fail in a DAQ reading testing ?

I want to write the while loop to read the DAQ signal. Each time, the DAQ input signal is compared with the spec. limit. The signal that is within the limit will be recorded as " Pass " while the signal out of spec. will be recorded as " Fail ". At the end of the testing ( I break the program), the testing result will be displayed on the testing panel.
I believe my question is similar to the programming of " A=A+1 ". Can anybody help me to resolve the problem ?

Tian,
Are you looking for just the number of passed and number failed measurements or a record of each individual test. If you are only concerned with the number of passed/failed tests then use a shift register. Right-click on the border of the loop and choose Add Shift Register. Then use a case structure. If it passed the test (true case) then increment the value. If it did not pass the test (false case) then just wire the value straight through the case structure. To get the number that failed simply subtract the number that passed from the iteration count.

Similar Messages

  • Why does captivate 5 scorm 2004 change the pass/fail status when using complete/incomplete

    SUMMARY:  I am trying to use SCORM 2004 reporting but I am finding out ADOBE does not diffentiate between COMPLETE/INCOMPLETED and PASSED/FAILED variables in the SCORM 2004.
    RESULT: I can not accurately get the passing/failing status from the course.
    QUESTION: Is their anything I can do or does ADOBE have a patch they have release which will accurately report the passed/failed STATUS as well as the completed/incompleted status.
    OTHER RESEARCH: I even attempted to post these "proofs" on scorm.com and they have something called "SATISFACTION TRACKED" to accomodate content creation systems (CAPTIVATE 5) which does not follow the standards of SCORM 2004.
    BACKGROUND:  the report status "Incomplete/Complete" and "Pass/Fail" under the publishing options for SCORM 2004 essentially mean the same thing when the CAPTIVATE COURSE sends setvar commands to the SCORM API...  In other words the SCORM API receives the same exact commands when publishing in both modes.  "THIS IS WRONG"
    why is the wrong?  IF REPORTING; incomplete/complete ;  the CAPTIVATE 5 COURSE SHOULD NOT "change" the STATUS OF PASS/FAIL when attempting to report COMPLETE/INCOMLETE with 100% slide views.
    DEMO #3> SHOWS THE PROBLEM
    About DEMO #1 and DEMO #2> these are other reporting options I attempted with other problems arrising..so these images are just for supporting my original problem which is shown in DEMO #3...
    DEMO 1 (REPORT STATUS has no affect on reporting):SETTINGS:
    PASS: (PROBLEM)
    Anything above Passing Score is not recorded to LMS (once passing is reached...no more bookmarks are recorded)
    FAIL: (CORRECT)
    Failure gets sent to the database on the last screen with "completed"
    DEMO 2 (Report status  [Complete/Incomplete  Pass/Fail]   has no impact on commands sent to scorm api):
    SETTINGS:
    PASS: (CORRECT)
    Very Last page gets sent to the Database
    FAIL: (PROBLEM)
    NOTHING GETS SENT TO DATABASE (stuck in final score screen forever)...because it says "INCOMPLETE"
    DEMO 3:
    Report status [Complete/Incomplete Pass/Fail] has no impact on commands sent to scorm api
    SETTINGS:
    PASS: (CORRECT)
    Very Last page gets sent to the Database with a 90% passing with "completed"
    FAIL: (PROBLEM)
    LAST SLIDE SENDS "completed" but GETS SENT AS A PASSING SCORE??? even when it is a failure
    (IN OTHER WORDS ADOBE CHANGES THE PASS/FAIL STATUS based on slide views...
    this is WRONG ....REALLY WRONG becuase I am trying to report
    "COMPLETE/INCOMPLETE" for 100% slide views
    PASS/FAIL OPTIONS:
    OPTION I CAN PURSUE?
    SO WHAT CAN I DO?  YES, I CAN TELL THE LMS...to COMPARE THE SCORE RECEIVED DURING A "COMPLETE" status...and then change the grade in the LMS based on the SCORE IN THE LMS...BUT THAT WOULD MEAN THE LMS COULD REPORT ONE GRADE and the COURSE COULD REPORT A DIFFERENT GRADE....and that would not be good.
    ALL OF OUR COURSES SHOW THE STUDENT THE FINAL SCORE WITHIN EACH COURSE..
    QUESTION TO ADOBE?
    SO ADOBE when are you going to differentiate these two reporting options so that the PASS/FAIL OPTIONS ARE ACCURATELY REPORTED WHEN USING COMPLETE/INCOMPLETE?
    MORE INFORMATION
    HERE IS THE SCORM.COM "Satisfaction" adjustment (very last line) to accomodate inconsistent reporting in CAPTIVATE 5.
    Thanks,
    DOUG LUBEY OF LOUISIANA
    www.douglubey.com
    OTHER WEBSITES:
    http://forums.adobe.com/message/3332035
    http://forums.adobe.com/message/3122315
    http://forums.adobe.com/thread/712502?start=50&tstart=0
    http://forums.adobe.com/thread/800570
    http://forums.adobe.com/thread/720121
    SEARCH ENGINE REFERENCE:
    CAPTIVATE SCORM SUCCESS_STATUS
    captivate 5 scorm 2004 pass/fail is dependent on complete/incomplete
    scorm 2004 captivate cs5 success_status incorrect
    scorm 2004 captivate cs5 reports success_status incorrectly
    scorm 2004 captivate incorrectly reports pass/fail
    scorm 2004 captivate 5 incorrectly reports pass/fail
    scorm 2004 "captivate 5" incorrectly reports pass/fail
    captivate 5 sends a pass even when a failure is issued by the course using scorm 2004.
    captivate does not work properly when working with completed status.
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS SHOULD BE INDEPENDENT
    CAPTIvATE 5 cmi.completion_status
    and cmi.success_status does not work independently with scorm 2004.  they both share the same definition.  These two field were designed to help differentiate the difference between a passing or failing score and when the course is actually completed.   WITH CAPTIVATE 5 they do not differentiate the two.  COMPLETE means PASS and INCOMPLETE means fail..
    I was trying to use "completion_status" as a way to determine when all slides were shown.  So that I can send the final score (PASS/FAIL) the LMS after all slides were viewed and this was the absoluted score the student recieved.
    2011 CAPTIVATE CHANGES cmi.success_status TO PASSING WHEN COURSE REPORTS FAILURE
    2010 CAPTIVATE CHANGES SUCCESS TO PASSED WHEN COURSE REPORTS FAILURE
    CAPTIVATE 5 CHANGES cmi.success_status TO PASSING
    CAPTIVATE 5 CHANGES cmi.success_status TO PASSING WHEN COURSE REPORTS FAILURE
    CPTX CAPTIVATE CHANGES cmi.success_status TO PASSING
    CP5 WHY DOES CAPTIVATE SCORM SUCCESS_STATUS
    CPTX WHY DOES CAPTIVATE CHANGE PASSING STATUS
    CPTX CAPTIVATE SCORM CHANGES FAILING SCORE TO PASSING SCORE
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS MEANS THE SAME THING
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS SHOULD BE INDEPENDENT

    Hi,
    Here's something I found in a similar situation... hope this helps!
    After publishing your CP5 project as a SCORM 2004 zip file, unzip the file so you can get into the separate files.  Open the imsmanifest.xml file with a text editor; I like using Notepad++.  Search for the line starting with <imsss:minNormalizedMeasure>.  The complete line will probably look like <imss:minNormalizedMeasure></imss:minNormalizedMeasure>, this is the line that is supposed to contain your passing score (at least for our LMS).  So, if your passing score is supposed to be 80%, edit this line to be:  <imsss:minNormalizedMeasure>0.80</imsss:minNormalizedMeasure> and save the file.  Then select all the files and re-zip them, I use 7-zip for that.
    I found one other weird thing that doesn't happen 100% of the time, so test it.  In the example above, try to get a score of exactly 80% (or whatever passing grade you are using) and see if you get the correct status reported.  I found I sometimes have to re-edit that line in the manifest file to be 0.79 so it correctly includes an exact 80% score.  No idea why and you have to watch the number of questions in your quiz to make sure that works for you mathematically.
    If this works or not, I'd recommend posting your results so others can get some ideas.  Good Luck!!

  • Can the sequence pass/fail status be controlled in Test Stand?

    I have a sequence that contains a DoWhile loop. Within the loop are 5 or 6 steps that are executed. If the result of those steps is a fail the loop runs again until it passes or meets its allowable number of iterations. My problem is that if an iteration fails and then passes the result for the sequence is failed. Can the pass/fail status be controlled via an expression etc? I know I can create a subsequence and loop in it with the properties. I am just curious to know if there is a way it can be done like I have explained.
    Thanks in advance for any help,
    Troy

    Troy,
    You might be able to use the RunState.ThisContext.SequenceFailed property.  Set this equal to False after your loop runs, and the sequence should report as if it passed.
    John B.
    Applications Engineer
    National Instruments

  • Combine "Pass/Fail" and "Multiple numeric limet" tests

    Hi,
    how is it possible to combine a pass/fail and a multiple numeric limit test?
    I know that you can create your own step types and so I already combined these two steps. But now i can only observe the numeric limits. Regardless of the pass/fail source ("Edit Pass/Fail Source") TestStand only looks at the numeric limits. If they are ok the test is passed. If not, it is failed. It does not matter if "Step.Result.PassFail" is True or False.
    What is the problem, or do i make any mistakes?
    Thanks for your help
    Daniel

    Hi,
    You will need to include the Step.Result.PassFail in the Status Expression. Probably at the moment you only have the NumericLimitTest Status expression in your combined step type. You will need to changed the Status Expression to match your requirement which will give you an overall Pass Fail for your step type.
    Regards
    Ray Farmer
    Regards
    Ray Farmer

  • Trouble with the pass/fail test

    I started having this trouble just recently and I can't figure out the problem. Just all of a sudden (without me changing anything anywhere apparently), whenever I call a pass/fail test and if the test fails, the sequence execution immediately goes to the <End Group> step. I want the execution to proceed with the next step regardless of a pass or a fail. I have rechecked the run options repeatedly and there's nothing in there that might be causing this. Also, if a pass/fail step fails and I am running on single pass, even the report does not get generated.
    I know it can be really hard for someone out there to tell me what the problem is. I am posting this message here just to see if people have any general comments/suggestions on what might be going on (perhaps this is a common problem which someone has run into before). The comments/suggestions might help me debug the problem.
    Thanks
    Anuj

    Hi Anuj,
    Do you have other step type in the same sequence or just Pass/Fail type?
    What adapter type are you using, LabVIEW, CVI, etc...
    Does it happen in the just one sequence or any sequence?
    What versions are you using?
    Posting an example would help.
    Regard
    Ray Farmer
    Regards
    Ray Farmer

  • What is property that contains the pass/fail status(UUT Result) of my main sequence ?

    Test Engineering
    Empower RF Systems Inc.

    Hi Mal,
    In general, in TestStand you use a step type that can use the Sequence Adapter in order to call a sequence (named subsequence in this context). The MainSequence is called from within the ProcessModel execution entry-points ('Test UUTs' and 'Single Pass' sequences in the TestStandModel.seq) when you launch the execution from "Execute->Test UUTs" or "Execute->Single Pass" menu or directly by the engine when you start an Interactive execution by selecting "Execute->Run MainSequence" menu option.
    In the first case, after the sequence call step executes, the Sequence Adapter can set the step status. So, finally the property that contains the status of Main Sequence is a subproperty of the SequenceCall step that in the ProcessModel execution entry-points is named "MainSeque
    nce Callback". The lookup string for this property depends from where you want to access it. For example :
    a) if it's from a step you add as part of the Execution entry-points customization, then the lookup string for the property should be "RunState.Sequence.Main["MainSequence Callback"].Result.Status"
    b) if it's from a step in the MainSequence, then the lookup string for the property should be "RunState.Caller.RunState.Sequence.Main["MainSequence Callback"].Result.Status" but special consideration should be taken if the MainSequence is run in Interactive mode, since the "RunState.Caller" property doesn't exist in this case.
    In the second case the sequence status isn't available (see above).
    Regards,
    Silvius
    Silvius Iancu

  • I am trying to use the T. S. operator interface I need to add pass fail and error

    Hello:
    I  am using Labview 6.1 and Test Stand 2.0
    I have made some modification to a copy of the T.S. Labview operator interface. How do I get the pass/fail or error from test stand to be pass to LabView so I can tell the operator that the unit pass or fail the test. I do not want to use the Test UUTs. I want this to be very low operator response.
    Message Edited by Toro on 02-28-2006 01:52 PM

    Not sure if this is available in the earlier versions of TestStand, but I think so.
    You could post a UI Message, from your execution entry point in the process model, to your User Interface indicating the pass/fail/error status when a test run is complete.  You will need to handle the UI Message in your Operator Interface.  See TestStand II Customisation course notes for handling UI Messages.
    The data passed in the UI Message can be anything you like, the UI Message handler in the User Interface will have to process the message and update the display as required.
    Hope this helps
    Steve
    Message Edited by SercoSteve on 03-01-2006 10:02 AM
    There are 10 types of people in the world that understand binary, those that do and those that don't.

  • How do i configure a FOR loop to have the behavior of the step Loop Type: Pass/Fail count?

    Hello,
    I'm using the Pass/Fail count set to one Pass to capture an event generated by my DUT.  I originally used a numerical compare step with the Looping type of Pass/Fail count to accomplish this.  Unfortunately the implementation changed and now I need to execute a few steps that can not be combined within one code module as before. Nor can these steps be put into a subroutine.  One of the steps executes a .NET asembly and I haven't figured out how to pass the reference to the subroutine.  When the subroutine is intered the reference is lost and the methode does not execute correctly.
    I have an evaluation function the exits the loop when the expected conditions are met. Everything works except for the Overall Pass/Fail result of the For loop.  If the loop exits due to the first numerical compare test passing, I want the loop overall execution to report as "Passed".  If the loop reaches it's predetermined number of iterations, the overall result needs to report as "Failed".  It would also be nice to have the radio button functionality of "Record Result of Each iteration".  Some conditions require a wait over a minute for the event to occur and I don't want to generate needless data for the report.
    Currently I get the pass/fail status for each For loop iteration, but only "Done" for each loop iteration.  I don't want the initial few failures to cause the test to fail.
    Does anyone know how to do this? Suggestions?
    Thanks,

    I have 2 steps in the loop that can not be combined into one step. This forces me to implement the behavior in some form of loop.  A While Loop could be an option.  I would need to pass the Step.Result.Status from the Multiple Numerical compaire step to the condition of the While Loop.  I tried to use the Step.TS.ID without success.  Some form of reference to the numerical compare step would be needed. I'm using a For Loop as I do want to limit the number of iterations of the loop.  In the case where the loop iterations are reached, the event I am trying to detect did not occur at the correct time and a failure needs to be reported.
    I came up with something based on my comments in the second post:
    1) To start with I configured a Locals.ForLoop_5 variable.  This is used to set the limit on the loop iterations and for comparison after the loop has finished executing. More on that later.
    2) The first step inside the loop invokes a method within a .NET assembly that has been packed in a DLL This method gets the required data from the DUT and stores it to a text file.
    3) The next step is the Multiple Numeric Limit step.  This step invokes a VI that extracts the data from the text file.  This step has been customized in several ways.
      i)  In Run Options, the Results Recording Option was disabled.  This prevents recording of "Failed" while the loop executes when waiting for the event to happen.
      ii) In Run Options, the Step Failure Causes Sequence Failure is unchecked.  Same reasoning as i)  These steps are not true failures.
      iii) A Post Action is configured to go to the nexxt step after the For Loop End step On Condition True with the logic of Step.Result.Status == "Passed".  This causes the loop to exit when the first "Passed" is encountered which corrolates with the event I'm trying to detect. On Conditon Fail remains set to default.
    4)  The step after the For Loop End is an expression step with everythin set to default except for the Status Expression logic set to: Locals.Loopindex < Locals.ForLoop_5 ? (Step.Result.Status = "Passed") : (Step.Result.Status = "Failed"). This step performs the overall Pass/Fail reporting for the For Loop.  If the number of loop iterations is less than the maximum it could have only gotten there by the previous logic triggered by the numerical compare passing, therefore "Passed".  If the loop index has reached the limit, then the event was not detected, therefore Failed.
    I have tested this work around with success, it just a pain to now have to implement this on my 40 some odd For Loops.
    If there is a better way, I'd still like to hear it.

  • Pass/Fail issue with Quiz and Click Buttons

    Got a complicated one for ya!
    I have a project that has about 100 quiz questions.  I want to add slides with an image for backgrownd, a voice over and a "Click Button" to be used like a Simulation.  If the "Click Button" is not clicked on, I want the slide to act like a Quiz and be marked as wrong.
    At the end of the Quiz I want the Pass/Fail function to be: If Pass: "Continue" then if Fail "Jump to Slide". This Failed Slide would jump to the last slide; the end of the project with a "Thank you".  If Passed, the project continues to display more information on other slides.
    The tests I have done the Pass/Fail seems to not be working.  If failed, the project continues just like the Passed even though the "Pass/Fail" options are marked correctly.
    What do you think?

    Also check the options under Quiz > Settings > Required.  It could be that you have it set to something other than Optional which might be restricting things.

  • Tracking Pass/Fail on LMS

    I am trying to track pass/fail on a Captivate project and I
    am having difficulties with the LMS. It is tracking as completed
    after just the first slide, before even taking the quiz. I have my
    quiz settings set to report "Quiz results only", Pass/Fail, Report
    to LMS as Percent, Reporting level is "Report Score", the Pass/Fail
    options are set to 75%.
    NOTE: I have allowed the user 2 attempts to pass. Could this
    be what's giving me trouble?

    Look at the Project > Advanced Interaction panel and see
    if you have
    score values setup throughout your lesson as you should.
    It sounds like perhaps you have a low number of scored
    interactions and
    that 75% is being met as soon as they answer the first
    question
    correctly? Compare the Advanced Interaction results to your
    Quiz
    Preferences > Pass/Fail options and see what shakes out...
    Erik
    BriarPatch1 wrote:
    > I am trying to track pass/fail on a Captivate project
    and I am having
    > difficulties with the LMS. It is tracking as completed
    after just the first
    > slide, before even taking the quiz. I have my quiz
    settings set to report
    > "Quiz results only", Pass/Fail, Report to LMS as
    Percent, Reporting level is
    > "Report Score", the Pass/Fail options are set to 75%.
    >
    > NOTE: I have allowed the user 2 attempts to pass. Could
    this be what's giving
    > me trouble?
    >
    Erik Lord
    http://www.capemedia.net
    Adobe Community Expert - eLearning
    http://www.adobe.com/communities/experts/
    http://www.awaretips.net -
    Authorware Tips!

  • Get overall Pass/Fail from custom User Interface

    Hi Guys
    I’m working on a custom user interface for operators and I would like to check and display the overall Pass/Fail result of the test at certain time intervals during execution.
    I added a sequence called ‘GetOverallResult’ to the ‘FrontEndCallbacks.seq’ which checks the Pass/Fail status of the sequence and I’m calling it form a SubVI within the User Interface written in LabVIEW (Check status.vi).
    The problem is that I’m getting the Pass/Fail status of the CallBack sequence not the test sequence I’m running.
    Could anyone please help me in this?
    Attachments:
    Check status.vi ‏17 KB
    FrontEndCallbacks.seq ‏5 KB

    Hi Attila,
    Have you instead tried reading the value directly from the main sequence? Have a look at the attached VI; pass the sequence context in and the Get Property Value VI will read the sequence's state.
    Is this suitable for what you had in mind?
    Regards,
    Tom
    Applications Engineering, NI UK
    Attachments:
    check.vi ‏10 KB

  • SCO Mastery Score vs Pass/Fail  Options

    Hi,
    Could someone explain to me the deference between setting a
    the SCO Mastery Score at say 80% and setting the pass/fail option
    to 80%?
    It seems in content we have created in the past that SCO
    mastery score is essential in stopping content going to completion
    and to transcript while the pass/fail setting doesn't act in this
    way. So this leads to the question, what is the point of the
    pass/fail setting?
    Any advice would be greatly appreciated as always.
    Ryan

    Ryan, from what I have experianced with this, the pass/fail
    option determines the score that Captivate uses to decide if it
    should display pass of fail slides, or the message on the results
    slide. the SCO Mastery Score tells your LMS what to use for pass or
    fail.
    Dave.

  • Captivate 5.5 Quiz Preferences Pass/Fail Actions

    Is there a good explanation anywhere of the Pass/Fail actions in Captivate, when one would want to use an action and which combinations of actions can or should be used?
    No matter what I choose for actions, for both passing and failing grade, Disable, Enable, Increment, and Deincrement are grayed out. If they are always grayed out, what are they used for?
    I looked this up in the Adobe Help menu, but as often happens, the information is too vague and high level in the Help menu for any real understanding.

    Hi Jay,
    I always execute an advanced actions for Pass/Fal actions, that will give you all possible statements (except Increment... because they are never available in advanced actions, one of the feature requests I have already entered several times). In this blog post I tried to explain the functionalities of question slides and the score slide, and those Pass/Fail actions are included in the blog post:
    Question Question Slides in Captivate
    Lilybiri

  • Dynamically specify TestStand pass/fail string

    I am performing tests utilizing multiple instruments. Each instrumemt may independently be in 'live' or 'simulated' mode. If any one instrument is in 'sim' mode, the test is considered in 'sim' mode. When in 'sim' mode the tests pass.
    My goal is to be able to look at the test report and determine if a specific test ran in 'sim' mode or live mode. My solution is to append the suffix '-sim' to the test status produced by TestStand. Thus the 'passed', 'failed' strings would become 'passed-sim', 'failed-sim' in 'sim' mode. These strings would have to be given to TestStand at the same time as the measurement value but before the test exited.
    Any thoughts or ideas on how I might accomplish the dynamic setting of the pass/fail string? I don't want this to be a permanent change, I want TestStand to revert back to the original pass/fail strings for the next sequential step.
    Hurst C.

    Hi Believer,
    What you seem to be trying to do is modify the results or a property of the results from the previous step. But unfortunately modify this property will only be affective if you were to re-run that step again eg step back one step. What you need to be modifying is the property that is in the Locals.ResultList. i.e Locals.ResultList[0].ReportText for the first step to be recorded in the sequence, Locals.ResultList[nth].ReportText for the nth step in the sequence to be recorded.
    By modifying 'RunState.Sequence.Main["Clock test (4.0V)"].Result.ReportText="FLAG"' you are actually modifying the runtime version of the step before its executed. Therefore, providing you do not over write the contents of ReportText within you step, this initial value will be seen in the result.
    You can check this by stepping into you step and checking the value of ReportText within the step.
    Hope this clarifies things
    Regards
    Ray Farmer
    Regards
    Ray Farmer

  • Sequence call with Pass/Fail Test always returns Passed result

    In TestStand 2.0, I use Sequence Adapter with Pass/Fail Test step to call a subsequence, I always receive a Passed result, even as I force fail a step in the subsequence and enabled the Step Failure Causes Sequence Failure option.
    Attachments:
    Example1.seq ‏17 KB

    Tan -
    The status expression for the Pass/Fail step type does not look to see if the status has been previously set by the code module in this case the sequence call.
    Step.DataSource != "Step.Result.PassFail" ?
    Step.Result.PassFail = Evaluate(Step.DataSource) :
    False, Step.Result.PassFail ? "Passed" : "Failed"
    It just set it to Passed or Failed even if the status is already "Failed".
    You could either,
    1) Change the Pass/Fail expression to the following for the step,
    Step.Result.Status == "Done"? Step.Result.PassFail : False
    or
    2) Use two steps, a Sequence Call and a Pass/Fail that uses the None Adapter. The Pass/Fail could operate on the previous step if necessary.
    Scott Richardson - NI"
    Scott Richardson
    National Instruments

Maybe you are looking for