Reporting results: Pass/Fail AND Incomplete

Some of our students may not complete a quiz and we want to
identify them so we can contact them. For those who do complete the
quiz, we need to report Pass or Fail. It doesn't look like we can
do both of these. Is that true? The settings seem to be radio
buttons that make you choose either/or.

I'd check into what you can do with LMS software. All of the
reporting needs above should be possible with the right Learning
Management System - which can be used to "plug in" your Captivate
SWF files. Or am I missing something here?

Similar Messages

  • [LV - TS] How to execute a SEQ file, get the result (PASS/FAIL) and the report file ?

    Hello,
    I'm looking for the most obvious and simple way to, from Labview :
    1- Start the Teststand engine : DONE
    2- Get a list of SEQ files to execute : DONE
    3- Loop through the list to execute each SEQ file : <- found some examples but I have some trouble with the UIMessage stuff
    http://zone.ni.com/reference/en-XX/help/370052N-01/tsexamples/infotopics/executing_sequences_using_api_labview/
    4- Get the result of the SEQ execution : <- unsure
    5- Get the XML path+name of the report file generated : <- obviously not easy
    http://digital.ni.com/public.nsf/allkb/3D3B2358DF9B786B8625789500549A3B
    http://forums.ni.com/t5/NI-TestStand/Getting-Report-File-Path-location-at-execution-in-TestStand-2012/td-p/3041789
    Here is my current implementation :
    David Koch
     

    Hello,
    I tried to work from this example :
    http://forums.ni.com/t5/NI-TestStand/Execution-of-TestStand-Sequence-in-LabVIEW-via-TS-API-Different/td-p/2939616/page/2#M46686
    However it remains unclear about how to get a simple
    answer from just running a sequence file out of any
    model so far (Simple Pass). Like I told you, all what
    I need, once the TS engine is started, is a VI with the
    following interface :
    INPUT
    1- TS engine reference
    2- SEQ file path (path or string)
    3- Sequence name to execute
    4- UUT string (optional, but yeah)
    5- error in
    OUTPUT
    1- Result boolean (PASS/FAIL)
    2- Path to XML report
    3- error out
    It would be so cool to just have this
    About the surrounding UI, I'm working on it. The LV
    code is not really important at this point, it's running
    a SEQ file that matters to me.
    David Koch
     

  • How can I copy test plan WITH ITS RESULTS (PASS/FAIL) to other project in MTM??

    How can I copy test plan WITH ITS RESULTS (PASS/FAIL) to other project in MTM??

    Hi odedn,
    Thank you for posting in MSDN forum.
    Based on your issue, you mean that you want to copy the test plan with test results from the current team project
    to other team project, am I right?
    If yes, as far as I know that when we copy test plan with test results to other team proejcts. It is default that
    there just will not copy the test plan with test results (pass/fail) to other team project in MTM.
    You can refer the following MSDN document’s
    what gets cloned, it explained that which objects are copied from the source test plan to the destination test plan.
    Reference:
    http://msdn.microsoft.com/en-us/library/hh543843.aspx
    If you still want to this feature, I suggest you can
    submit this feature request:
    http://visualstudio.uservoice.com/forums/121579-visual-studio. The Visual Studio product team is listening to user voice there. You can send your idea there and people can vote. If
    you submit this suggestion, I hope you could post that link here, I will help you vote it.
    Thanks for your understanding.
    Best Regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Combine "Pass/Fail" and "Multiple numeric limet" tests

    Hi,
    how is it possible to combine a pass/fail and a multiple numeric limit test?
    I know that you can create your own step types and so I already combined these two steps. But now i can only observe the numeric limits. Regardless of the pass/fail source ("Edit Pass/Fail Source") TestStand only looks at the numeric limits. If they are ok the test is passed. If not, it is failed. It does not matter if "Step.Result.PassFail" is True or False.
    What is the problem, or do i make any mistakes?
    Thanks for your help
    Daniel

    Hi,
    You will need to include the Step.Result.PassFail in the Status Expression. Probably at the moment you only have the NumericLimitTest Status expression in your combined step type. You will need to changed the Status Expression to match your requirement which will give you an overall Pass Fail for your step type.
    Regards
    Ray Farmer
    Regards
    Ray Farmer

  • Run Report with Pass Number and Varchar Lexical Parameter values

    Dear Sir/Madam
    Due to an urgent change required to a report I would most appreciate it if you can please advise me if it is possible to get this report working by either passing in multiple lexical parameters or one signle lexical parameter when calling the SRW.RUN_REPORT command with the following methods:
    Firstly: When calling report with a run_no and spr_cd passed in through one lexical:
    cmd_line: REPORT=D:\DEV\REMITTANCE.rdf BACKGROUND=YES BATCH=NO DESFORMAT=PDF DESTYPE=FILE DESNAME=D:\DEV\REMITTANCE.pdf
    CP_L_PARAM=RUN_NO=TO_NUMBER(28) AND SPR_CD=SUPP1
    Both with a TO_NUMBER and without, whereby I attempt to do a SUBSTR within the query of the called report to derive the RUN_NO Oracle Reports fails to accept the SUBSTR.
    Secondly: I am now trying to call the with passing two lexical parameters into the report as follows:
    cmd_line: REPORT=D:\DEV\REMITTANCE.rdf BACKGROUND=YES BATCH=NO DESFORMAT=PDF DESTYPE=FILE DESNAME=D:\DEV\REMITTANCE.pdf
    CP_L_RUN_NO=||TO_CHAR(lv_run_no)|| CP_L_SPR_CD=||lv_spr_cd;          
    NOTE: The above is an output of the actual command line and not what is passed into the command line, thus the quotes are missing. Please also note that the report is running fine with the only a hard coded RUN_NO value PASSED without the SPR_CD.
    Unfortunately this is also causing issues as the RUN_NO is a number and as you know you can only pass in strings.
    Your urgent help is required on this matter please as our client is expecting a solution this afternoon.
    Hope to hear form you soon.
    Kind regards
    Andrew Mason

    Dont Worry I've worked it out...

  • Edit the result (pass/fail) of a subsequence

    I have a sequence that looks like this:
    MainSequence
    Step1
    Step2
    SubSequence
    DoWhile (Test1 fail and time has not exceed 30 sec)
      Test1
    End
    Step3
    The DoWhile loop run Test1 over and over until either Test1 passes or time has exceed 30 seconds. Usually, Test1 actually pass. So, on the steps screen, Test1 would actually be shown as "PASS", but when the SubSequence exits, the SubSequence would be shown as "Failed". I think this is because Test1 failed a few times before passing.
    How do I change it so that the SubSequence would be shown as "PASS" in the report as long as Test1 passed once?
    Thanks.
    Solved!
    Go to Solution.

    Hi Bladhart,
    Maybe this small example
    will solve your task,
    Regards
    Juergen
    =s=i=g=n=a=t=u=r=e= Click on the Star and see what happens :-) =s=i=g=n=a=t=u=r=e=
    Attachments:
    Loop.seq ‏7 KB

  • I am trying to use the T. S. operator interface I need to add pass fail and error

    Hello:
    I  am using Labview 6.1 and Test Stand 2.0
    I have made some modification to a copy of the T.S. Labview operator interface. How do I get the pass/fail or error from test stand to be pass to LabView so I can tell the operator that the unit pass or fail the test. I do not want to use the Test UUTs. I want this to be very low operator response.
    Message Edited by Toro on 02-28-2006 01:52 PM

    Not sure if this is available in the earlier versions of TestStand, but I think so.
    You could post a UI Message, from your execution entry point in the process model, to your User Interface indicating the pass/fail/error status when a test run is complete.  You will need to handle the UI Message in your Operator Interface.  See TestStand II Customisation course notes for handling UI Messages.
    The data passed in the UI Message can be anything you like, the UI Message handler in the User Interface will have to process the message and update the display as required.
    Hope this helps
    Steve
    Message Edited by SercoSteve on 03-01-2006 10:02 AM
    There are 10 types of people in the world that understand binary, those that do and those that don't.

  • TestStand Lookup String for UUT Result Passed/Failed

    When I am in CleanUp group,what is the lookup string for UUT Result=Passed or Failed. That is all steps in main group passed = UUT Result=Passed. I am using Labview.

    Using RunState.SequenceFailed called from Cleanup step in TestStand sequence seems to work reliably for me but using RunState. ErrorReported does not. It seems as though the sequence context becomes invalid or something (although I get no error when accessing this property). It does not always come back true when a runtime error has occurred. Any ideas?
    Thanks, Roger Owens [email protected]

  • How can we send status to the LMS other than pass/fail, complete/incomplete?

    We have a Captivate module generated as SCORM 1.2 which passes Complete/Incomplete to the LMS. The quiz is at the end of the module - if the user exits the module prior to attempting the quiz, we would like to send 'Not Evaluated' back to the LMS rather than Incomplete (which happens today).
    Module is developed using Captivate 7.0.1
    Does anyone have a solution for this problem?

    Hi,
    You should do it using the FM for FTP´ing the files, check the link bellow:
    http://www.members.tripod.com/abap4/FTP_Using_SAP_Functions.html
    Hope this helps!!!
    Gabriel P.

  • Why does captivate 5 scorm 2004 change the pass/fail status when using complete/incomplete

    SUMMARY:  I am trying to use SCORM 2004 reporting but I am finding out ADOBE does not diffentiate between COMPLETE/INCOMPLETED and PASSED/FAILED variables in the SCORM 2004.
    RESULT: I can not accurately get the passing/failing status from the course.
    QUESTION: Is their anything I can do or does ADOBE have a patch they have release which will accurately report the passed/failed STATUS as well as the completed/incompleted status.
    OTHER RESEARCH: I even attempted to post these "proofs" on scorm.com and they have something called "SATISFACTION TRACKED" to accomodate content creation systems (CAPTIVATE 5) which does not follow the standards of SCORM 2004.
    BACKGROUND:  the report status "Incomplete/Complete" and "Pass/Fail" under the publishing options for SCORM 2004 essentially mean the same thing when the CAPTIVATE COURSE sends setvar commands to the SCORM API...  In other words the SCORM API receives the same exact commands when publishing in both modes.  "THIS IS WRONG"
    why is the wrong?  IF REPORTING; incomplete/complete ;  the CAPTIVATE 5 COURSE SHOULD NOT "change" the STATUS OF PASS/FAIL when attempting to report COMPLETE/INCOMLETE with 100% slide views.
    DEMO #3> SHOWS THE PROBLEM
    About DEMO #1 and DEMO #2> these are other reporting options I attempted with other problems arrising..so these images are just for supporting my original problem which is shown in DEMO #3...
    DEMO 1 (REPORT STATUS has no affect on reporting):SETTINGS:
    PASS: (PROBLEM)
    Anything above Passing Score is not recorded to LMS (once passing is reached...no more bookmarks are recorded)
    FAIL: (CORRECT)
    Failure gets sent to the database on the last screen with "completed"
    DEMO 2 (Report status  [Complete/Incomplete  Pass/Fail]   has no impact on commands sent to scorm api):
    SETTINGS:
    PASS: (CORRECT)
    Very Last page gets sent to the Database
    FAIL: (PROBLEM)
    NOTHING GETS SENT TO DATABASE (stuck in final score screen forever)...because it says "INCOMPLETE"
    DEMO 3:
    Report status [Complete/Incomplete Pass/Fail] has no impact on commands sent to scorm api
    SETTINGS:
    PASS: (CORRECT)
    Very Last page gets sent to the Database with a 90% passing with "completed"
    FAIL: (PROBLEM)
    LAST SLIDE SENDS "completed" but GETS SENT AS A PASSING SCORE??? even when it is a failure
    (IN OTHER WORDS ADOBE CHANGES THE PASS/FAIL STATUS based on slide views...
    this is WRONG ....REALLY WRONG becuase I am trying to report
    "COMPLETE/INCOMPLETE" for 100% slide views
    PASS/FAIL OPTIONS:
    OPTION I CAN PURSUE?
    SO WHAT CAN I DO?  YES, I CAN TELL THE LMS...to COMPARE THE SCORE RECEIVED DURING A "COMPLETE" status...and then change the grade in the LMS based on the SCORE IN THE LMS...BUT THAT WOULD MEAN THE LMS COULD REPORT ONE GRADE and the COURSE COULD REPORT A DIFFERENT GRADE....and that would not be good.
    ALL OF OUR COURSES SHOW THE STUDENT THE FINAL SCORE WITHIN EACH COURSE..
    QUESTION TO ADOBE?
    SO ADOBE when are you going to differentiate these two reporting options so that the PASS/FAIL OPTIONS ARE ACCURATELY REPORTED WHEN USING COMPLETE/INCOMPLETE?
    MORE INFORMATION
    HERE IS THE SCORM.COM "Satisfaction" adjustment (very last line) to accomodate inconsistent reporting in CAPTIVATE 5.
    Thanks,
    DOUG LUBEY OF LOUISIANA
    www.douglubey.com
    OTHER WEBSITES:
    http://forums.adobe.com/message/3332035
    http://forums.adobe.com/message/3122315
    http://forums.adobe.com/thread/712502?start=50&tstart=0
    http://forums.adobe.com/thread/800570
    http://forums.adobe.com/thread/720121
    SEARCH ENGINE REFERENCE:
    CAPTIVATE SCORM SUCCESS_STATUS
    captivate 5 scorm 2004 pass/fail is dependent on complete/incomplete
    scorm 2004 captivate cs5 success_status incorrect
    scorm 2004 captivate cs5 reports success_status incorrectly
    scorm 2004 captivate incorrectly reports pass/fail
    scorm 2004 captivate 5 incorrectly reports pass/fail
    scorm 2004 "captivate 5" incorrectly reports pass/fail
    captivate 5 sends a pass even when a failure is issued by the course using scorm 2004.
    captivate does not work properly when working with completed status.
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS SHOULD BE INDEPENDENT
    CAPTIvATE 5 cmi.completion_status
    and cmi.success_status does not work independently with scorm 2004.  they both share the same definition.  These two field were designed to help differentiate the difference between a passing or failing score and when the course is actually completed.   WITH CAPTIVATE 5 they do not differentiate the two.  COMPLETE means PASS and INCOMPLETE means fail..
    I was trying to use "completion_status" as a way to determine when all slides were shown.  So that I can send the final score (PASS/FAIL) the LMS after all slides were viewed and this was the absoluted score the student recieved.
    2011 CAPTIVATE CHANGES cmi.success_status TO PASSING WHEN COURSE REPORTS FAILURE
    2010 CAPTIVATE CHANGES SUCCESS TO PASSED WHEN COURSE REPORTS FAILURE
    CAPTIVATE 5 CHANGES cmi.success_status TO PASSING
    CAPTIVATE 5 CHANGES cmi.success_status TO PASSING WHEN COURSE REPORTS FAILURE
    CPTX CAPTIVATE CHANGES cmi.success_status TO PASSING
    CP5 WHY DOES CAPTIVATE SCORM SUCCESS_STATUS
    CPTX WHY DOES CAPTIVATE CHANGE PASSING STATUS
    CPTX CAPTIVATE SCORM CHANGES FAILING SCORE TO PASSING SCORE
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS MEANS THE SAME THING
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS SHOULD BE INDEPENDENT

    Hi,
    Here's something I found in a similar situation... hope this helps!
    After publishing your CP5 project as a SCORM 2004 zip file, unzip the file so you can get into the separate files.  Open the imsmanifest.xml file with a text editor; I like using Notepad++.  Search for the line starting with <imsss:minNormalizedMeasure>.  The complete line will probably look like <imss:minNormalizedMeasure></imss:minNormalizedMeasure>, this is the line that is supposed to contain your passing score (at least for our LMS).  So, if your passing score is supposed to be 80%, edit this line to be:  <imsss:minNormalizedMeasure>0.80</imsss:minNormalizedMeasure> and save the file.  Then select all the files and re-zip them, I use 7-zip for that.
    I found one other weird thing that doesn't happen 100% of the time, so test it.  In the example above, try to get a score of exactly 80% (or whatever passing grade you are using) and see if you get the correct status reported.  I found I sometimes have to re-edit that line in the manifest file to be 0.79 so it correctly includes an exact 80% score.  No idea why and you have to watch the number of questions in your quiz to make sure that works for you mathematically.
    If this works or not, I'd recommend posting your results so others can get some ideas.  Good Luck!!

  • Can I build a test over several pages, and give instant pass/fail feedback.

    I want to build a test over several pages.  I would like to have buttons where you click on the answer, and at the end of the test show pass/fail and answers to failed questions.  I also want to show the percentage to the test answers, such as 76% out of 100%.  I would like to adjust the percentages depending on the  amount of answers per questions.  Let's say there are 7 questions for the test, a passing score would be 71%, if there are 5 questions, a passing score would be 80% and so forth.
    I am sure there are many questions regarding this, feel free to ask any and I will answer as much as possible.
    I have Keynote 09, Numbers 09 and Pages 09, I would prefer to work with Keynote and hopefully use  the other  two as needed.

    chrisfrommarion wrote:
    Thanks, while it worked on yours it doesn't work on mine, any other ideas?
    Are you holding down the "d" key, all by itself, just after the system shuts down for a restart?
    And how long did you hold it down?
    I held mine down for about 40-50 seconds. Then I lookat at my routers WAN light and it was flickering so I let up on the d key.
    If that does not work then maybe there is something wrong with your MBP. Is that the reason you want to run this test, something isn't right with your unit?

  • SCORM - Simple Quiz - Pass/Fail Only

    Flash CS4
    AS2
    Good Afternoon,
    I've created a simple multiple choice quiz in Flash.
    I was wondering if anyone could point me to some resources for writing a simple SCORM wrapper, and the accompanying AS2 code.  This is only going to report a pass/fail to the LMS - that's all - nothing more at this point.  I'm hoping there should be a simple tut on this, as I said, it's pretty straight forward.
    Thank you!
    Actus

    I have a tutorial here:
    http://pipwerks.com/2008/04/27/how-to-add-basic-scorm-code-to-a-flash-movie/
    It uses a custom (and free) SCORM ActionScript class and JavaScript wrapper so you don't need to write your own.
    The tutorial uses SCORM 2004, but the JavaScript wrapper and ActionScript classes are version-neutral (works for SCORM 2004 and 1.2). There are two versions of the ActionScript class available: one for AS2 and one for AS3.
    You can download the wrapper files here:
    http://pipwerks.com/downloads/
    - philip

  • Reports Background Engine fails

    Hello,
    sometimes happens that Reports Background Engine fail and Form application stop responding. Only way is shutdown Form application. This problem occur randomly, but on some computers it happens frequently with the same report.
    Where is the bug?
    Thanx

    pls contact oracle support services for further investigation of your problem. this requires detailed investigation of your environment that goes beyond what can be done here in the forum.
    thanks,
    ph.

  • How to change the result status "Failed" after the test to "Passed"

    We make a picture of a LCD screen and look if its all working. Sometimes the LCD screen is working propper but the Step Failed. In case automatic LCD test fails we show the picture from the camera to the user and he can decide if the LCD screen passed or failed. So when the user decide that the screen is working propper we should overwrite the status of the automatic test that in the report and database shows lcd screen test passed. I tried with 
    RunState.SequenceFile.Data.Seq["Check_LCD_V3-0"].M​ain["154 Execute - All Horizontal Segments On"].Result.Status="Passed",
    RunState.SequenceFile.Data.Seq["Check_LCD_V3-0"].M​ain["154 Execute - All Horizontal Segments On"].Result.PassFail="True",
    But in the database the Status of this step is still "Failed"
    What i do wrong?
    Thanks
    Solved!
    Go to Solution.

    Alteschwed wrote:
    We make a picture of a LCD screen and look if its all working. Sometimes the LCD screen is working propper but the Step Failed. In case automatic LCD test fails we show the picture from the camera to the user and he can decide if the LCD screen passed or failed. So when the user decide that the screen is working propper we should overwrite the status of the automatic test that in the report and database shows lcd screen test passed. I tried with 
    RunState.SequenceFile.Data.Seq["Check_LCD_V3-0"].M​ain["154 Execute - All Horizontal Segments On"].Result.Status="Passed",
    RunState.SequenceFile.Data.Seq["Check_LCD_V3-0"].M​ain["154 Execute - All Horizontal Segments On"].Result.PassFail="True",
    But in the database the Status of this step is still "Failed"
    What i do wrong?
    Thanks
    I recommend the following:
    1) Make the automated LCD test an Action step rather than a Pass/Fail step and instead set a local variable to whether the automatic LCD test failed (e.g. Locals.AutomatedLCDFailed).
    2) Create a new Pass/Fail test (using the none adapter) that takes both the local variable and the user's input into account to determine it's pass/fail status (e.g. Set the step's data source to an expression like: !(Locals.AutomatedLCDTestFailed && Locals.UserLCDTestFailed)
    As an alternative to making 1) an Action step you could leave it a pass/fail step and uncheck the step setting on it for "Step failure causes sequence failure". Then you could use the status of that step instead of Locals.AutomatedLCDTestFailed. The one difference is that the automated test will still show up as failed in your report even though the sequence can still be passed. It's somewhat a matter of preference.
    Hope this helps,
    -Doug

  • Dynamically passing text and url-based images as an input parameter to cf8 report builder

    I'm unsuccessfully trying to dynamically pass text and url-based images to a group footer or the detail section via an input parameter or even hardcoded. The field has the attribute 'XHTML Text Formating' set to True. The following are failed samples of a simplified value:
    "<img height=’300’ alt=’Document’ width=’300’ src=’http://www.google.com/intl/en_ALL/images/logo.gif’ />"
    or
    "<img src=’http://www.google.com/intl/en_ALL/images/logo.gif’ />"
    This just results in the above text being output. The end result would have various text and images from a database as input by a user, thus the reason I cannot just use the hyperlink information attribute as I could if it were a single known image. I tried rtf and pdf report types. Ideas?

    HTH,
    Thanks. I'll keep that in mind, although I don't know how many images my user might need or what sizes so that might be tricky.
    Since my target output is rtf so that MS Word can be used to edit the result, I added a pagebreak to a MS Word doc and used the resulting html source to replace the rich text editor source code for the page break, but that did not help either. The page break was so a user could add an image later. Something is wrong with the Report Builder related to intepreting XHTML, especially anything that has an attribute, including URL-based image links. I hope they try to provide another update before CF9. I doubt my client will be going to CF9 for some time, since they are just completing the migration to CF8.
    BrianO

Maybe you are looking for