Pass/Fail Assessments

I am using Captivate 3 and am looking for resources that can guide me through creating an assessment in a pass/fail mode.  I have created assessments, and have created quizzes, but need to bring the two together to record a demonstration of a software package we use, then launch an assessment that requires the user to get XX percent of mouse clicks, etc. correct in order to pass the assessment.  I know that I can use failure captions and set the number of tries a user is allowed, but have not been able to determine a way to set it up as pass/fail.
Thanks for your assistance.

Hello,
No problem with CP. First you will have to determine the mastery score, what is needed to Pass. This has to be done in the Quiz, Preferences, Pass or Fail, Pass options. You can combine Question slides with assessment, where the user has to use interactive objects to be scored. An overview of scored Question slides and interactive objects can be seen in the view 'Advanced Interaction' (menu Project). Then you will have to set the Reporting (also in Quiz Preferences), depending on the SCORM version (do not know Plateau myself). It is possible to report Pass/fail as you will see in those settings.
Hope this can be a start.
Lilybiri

Similar Messages

  • I am trying to use the T. S. operator interface I need to add pass fail and error

    Hello:
    I  am using Labview 6.1 and Test Stand 2.0
    I have made some modification to a copy of the T.S. Labview operator interface. How do I get the pass/fail or error from test stand to be pass to LabView so I can tell the operator that the unit pass or fail the test. I do not want to use the Test UUTs. I want this to be very low operator response.
    Message Edited by Toro on 02-28-2006 01:52 PM

    Not sure if this is available in the earlier versions of TestStand, but I think so.
    You could post a UI Message, from your execution entry point in the process model, to your User Interface indicating the pass/fail/error status when a test run is complete.  You will need to handle the UI Message in your Operator Interface.  See TestStand II Customisation course notes for handling UI Messages.
    The data passed in the UI Message can be anything you like, the UI Message handler in the User Interface will have to process the message and update the display as required.
    Hope this helps
    Steve
    Message Edited by SercoSteve on 03-01-2006 10:02 AM
    There are 10 types of people in the world that understand binary, those that do and those that don't.

  • How do i configure a FOR loop to have the behavior of the step Loop Type: Pass/Fail count?

    Hello,
    I'm using the Pass/Fail count set to one Pass to capture an event generated by my DUT.  I originally used a numerical compare step with the Looping type of Pass/Fail count to accomplish this.  Unfortunately the implementation changed and now I need to execute a few steps that can not be combined within one code module as before. Nor can these steps be put into a subroutine.  One of the steps executes a .NET asembly and I haven't figured out how to pass the reference to the subroutine.  When the subroutine is intered the reference is lost and the methode does not execute correctly.
    I have an evaluation function the exits the loop when the expected conditions are met. Everything works except for the Overall Pass/Fail result of the For loop.  If the loop exits due to the first numerical compare test passing, I want the loop overall execution to report as "Passed".  If the loop reaches it's predetermined number of iterations, the overall result needs to report as "Failed".  It would also be nice to have the radio button functionality of "Record Result of Each iteration".  Some conditions require a wait over a minute for the event to occur and I don't want to generate needless data for the report.
    Currently I get the pass/fail status for each For loop iteration, but only "Done" for each loop iteration.  I don't want the initial few failures to cause the test to fail.
    Does anyone know how to do this? Suggestions?
    Thanks,

    I have 2 steps in the loop that can not be combined into one step. This forces me to implement the behavior in some form of loop.  A While Loop could be an option.  I would need to pass the Step.Result.Status from the Multiple Numerical compaire step to the condition of the While Loop.  I tried to use the Step.TS.ID without success.  Some form of reference to the numerical compare step would be needed. I'm using a For Loop as I do want to limit the number of iterations of the loop.  In the case where the loop iterations are reached, the event I am trying to detect did not occur at the correct time and a failure needs to be reported.
    I came up with something based on my comments in the second post:
    1) To start with I configured a Locals.ForLoop_5 variable.  This is used to set the limit on the loop iterations and for comparison after the loop has finished executing. More on that later.
    2) The first step inside the loop invokes a method within a .NET assembly that has been packed in a DLL This method gets the required data from the DUT and stores it to a text file.
    3) The next step is the Multiple Numeric Limit step.  This step invokes a VI that extracts the data from the text file.  This step has been customized in several ways.
      i)  In Run Options, the Results Recording Option was disabled.  This prevents recording of "Failed" while the loop executes when waiting for the event to happen.
      ii) In Run Options, the Step Failure Causes Sequence Failure is unchecked.  Same reasoning as i)  These steps are not true failures.
      iii) A Post Action is configured to go to the nexxt step after the For Loop End step On Condition True with the logic of Step.Result.Status == "Passed".  This causes the loop to exit when the first "Passed" is encountered which corrolates with the event I'm trying to detect. On Conditon Fail remains set to default.
    4)  The step after the For Loop End is an expression step with everythin set to default except for the Status Expression logic set to: Locals.Loopindex < Locals.ForLoop_5 ? (Step.Result.Status = "Passed") : (Step.Result.Status = "Failed"). This step performs the overall Pass/Fail reporting for the For Loop.  If the number of loop iterations is less than the maximum it could have only gotten there by the previous logic triggered by the numerical compare passing, therefore "Passed".  If the loop index has reached the limit, then the event was not detected, therefore Failed.
    I have tested this work around with success, it just a pain to now have to implement this on my 40 some odd For Loops.
    If there is a better way, I'd still like to hear it.

  • Why does captivate 5 scorm 2004 change the pass/fail status when using complete/incomplete

    SUMMARY:  I am trying to use SCORM 2004 reporting but I am finding out ADOBE does not diffentiate between COMPLETE/INCOMPLETED and PASSED/FAILED variables in the SCORM 2004.
    RESULT: I can not accurately get the passing/failing status from the course.
    QUESTION: Is their anything I can do or does ADOBE have a patch they have release which will accurately report the passed/failed STATUS as well as the completed/incompleted status.
    OTHER RESEARCH: I even attempted to post these "proofs" on scorm.com and they have something called "SATISFACTION TRACKED" to accomodate content creation systems (CAPTIVATE 5) which does not follow the standards of SCORM 2004.
    BACKGROUND:  the report status "Incomplete/Complete" and "Pass/Fail" under the publishing options for SCORM 2004 essentially mean the same thing when the CAPTIVATE COURSE sends setvar commands to the SCORM API...  In other words the SCORM API receives the same exact commands when publishing in both modes.  "THIS IS WRONG"
    why is the wrong?  IF REPORTING; incomplete/complete ;  the CAPTIVATE 5 COURSE SHOULD NOT "change" the STATUS OF PASS/FAIL when attempting to report COMPLETE/INCOMLETE with 100% slide views.
    DEMO #3> SHOWS THE PROBLEM
    About DEMO #1 and DEMO #2> these are other reporting options I attempted with other problems arrising..so these images are just for supporting my original problem which is shown in DEMO #3...
    DEMO 1 (REPORT STATUS has no affect on reporting):SETTINGS:
    PASS: (PROBLEM)
    Anything above Passing Score is not recorded to LMS (once passing is reached...no more bookmarks are recorded)
    FAIL: (CORRECT)
    Failure gets sent to the database on the last screen with "completed"
    DEMO 2 (Report status  [Complete/Incomplete  Pass/Fail]   has no impact on commands sent to scorm api):
    SETTINGS:
    PASS: (CORRECT)
    Very Last page gets sent to the Database
    FAIL: (PROBLEM)
    NOTHING GETS SENT TO DATABASE (stuck in final score screen forever)...because it says "INCOMPLETE"
    DEMO 3:
    Report status [Complete/Incomplete Pass/Fail] has no impact on commands sent to scorm api
    SETTINGS:
    PASS: (CORRECT)
    Very Last page gets sent to the Database with a 90% passing with "completed"
    FAIL: (PROBLEM)
    LAST SLIDE SENDS "completed" but GETS SENT AS A PASSING SCORE??? even when it is a failure
    (IN OTHER WORDS ADOBE CHANGES THE PASS/FAIL STATUS based on slide views...
    this is WRONG ....REALLY WRONG becuase I am trying to report
    "COMPLETE/INCOMPLETE" for 100% slide views
    PASS/FAIL OPTIONS:
    OPTION I CAN PURSUE?
    SO WHAT CAN I DO?  YES, I CAN TELL THE LMS...to COMPARE THE SCORE RECEIVED DURING A "COMPLETE" status...and then change the grade in the LMS based on the SCORE IN THE LMS...BUT THAT WOULD MEAN THE LMS COULD REPORT ONE GRADE and the COURSE COULD REPORT A DIFFERENT GRADE....and that would not be good.
    ALL OF OUR COURSES SHOW THE STUDENT THE FINAL SCORE WITHIN EACH COURSE..
    QUESTION TO ADOBE?
    SO ADOBE when are you going to differentiate these two reporting options so that the PASS/FAIL OPTIONS ARE ACCURATELY REPORTED WHEN USING COMPLETE/INCOMPLETE?
    MORE INFORMATION
    HERE IS THE SCORM.COM "Satisfaction" adjustment (very last line) to accomodate inconsistent reporting in CAPTIVATE 5.
    Thanks,
    DOUG LUBEY OF LOUISIANA
    www.douglubey.com
    OTHER WEBSITES:
    http://forums.adobe.com/message/3332035
    http://forums.adobe.com/message/3122315
    http://forums.adobe.com/thread/712502?start=50&tstart=0
    http://forums.adobe.com/thread/800570
    http://forums.adobe.com/thread/720121
    SEARCH ENGINE REFERENCE:
    CAPTIVATE SCORM SUCCESS_STATUS
    captivate 5 scorm 2004 pass/fail is dependent on complete/incomplete
    scorm 2004 captivate cs5 success_status incorrect
    scorm 2004 captivate cs5 reports success_status incorrectly
    scorm 2004 captivate incorrectly reports pass/fail
    scorm 2004 captivate 5 incorrectly reports pass/fail
    scorm 2004 "captivate 5" incorrectly reports pass/fail
    captivate 5 sends a pass even when a failure is issued by the course using scorm 2004.
    captivate does not work properly when working with completed status.
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS SHOULD BE INDEPENDENT
    CAPTIvATE 5 cmi.completion_status
    and cmi.success_status does not work independently with scorm 2004.  they both share the same definition.  These two field were designed to help differentiate the difference between a passing or failing score and when the course is actually completed.   WITH CAPTIVATE 5 they do not differentiate the two.  COMPLETE means PASS and INCOMPLETE means fail..
    I was trying to use "completion_status" as a way to determine when all slides were shown.  So that I can send the final score (PASS/FAIL) the LMS after all slides were viewed and this was the absoluted score the student recieved.
    2011 CAPTIVATE CHANGES cmi.success_status TO PASSING WHEN COURSE REPORTS FAILURE
    2010 CAPTIVATE CHANGES SUCCESS TO PASSED WHEN COURSE REPORTS FAILURE
    CAPTIVATE 5 CHANGES cmi.success_status TO PASSING
    CAPTIVATE 5 CHANGES cmi.success_status TO PASSING WHEN COURSE REPORTS FAILURE
    CPTX CAPTIVATE CHANGES cmi.success_status TO PASSING
    CP5 WHY DOES CAPTIVATE SCORM SUCCESS_STATUS
    CPTX WHY DOES CAPTIVATE CHANGE PASSING STATUS
    CPTX CAPTIVATE SCORM CHANGES FAILING SCORE TO PASSING SCORE
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS MEANS THE SAME THING
    CAPTIVATE SCORM SUCCESS_STATUS COMPLETE STATUS SHOULD BE INDEPENDENT

    Hi,
    Here's something I found in a similar situation... hope this helps!
    After publishing your CP5 project as a SCORM 2004 zip file, unzip the file so you can get into the separate files.  Open the imsmanifest.xml file with a text editor; I like using Notepad++.  Search for the line starting with <imsss:minNormalizedMeasure>.  The complete line will probably look like <imss:minNormalizedMeasure></imss:minNormalizedMeasure>, this is the line that is supposed to contain your passing score (at least for our LMS).  So, if your passing score is supposed to be 80%, edit this line to be:  <imsss:minNormalizedMeasure>0.80</imsss:minNormalizedMeasure> and save the file.  Then select all the files and re-zip them, I use 7-zip for that.
    I found one other weird thing that doesn't happen 100% of the time, so test it.  In the example above, try to get a score of exactly 80% (or whatever passing grade you are using) and see if you get the correct status reported.  I found I sometimes have to re-edit that line in the manifest file to be 0.79 so it correctly includes an exact 80% score.  No idea why and you have to watch the number of questions in your quiz to make sure that works for you mathematically.
    If this works or not, I'd recommend posting your results so others can get some ideas.  Good Luck!!

  • How can I copy test plan WITH ITS RESULTS (PASS/FAIL) to other project in MTM??

    How can I copy test plan WITH ITS RESULTS (PASS/FAIL) to other project in MTM??

    Hi odedn,
    Thank you for posting in MSDN forum.
    Based on your issue, you mean that you want to copy the test plan with test results from the current team project
    to other team project, am I right?
    If yes, as far as I know that when we copy test plan with test results to other team proejcts. It is default that
    there just will not copy the test plan with test results (pass/fail) to other team project in MTM.
    You can refer the following MSDN document’s
    what gets cloned, it explained that which objects are copied from the source test plan to the destination test plan.
    Reference:
    http://msdn.microsoft.com/en-us/library/hh543843.aspx
    If you still want to this feature, I suggest you can
    submit this feature request:
    http://visualstudio.uservoice.com/forums/121579-visual-studio. The Visual Studio product team is listening to user voice there. You can send your idea there and people can vote. If
    you submit this suggestion, I hope you could post that link here, I will help you vote it.
    Thanks for your understanding.
    Best Regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Pass/Fail issue with Quiz and Click Buttons

    Got a complicated one for ya!
    I have a project that has about 100 quiz questions.  I want to add slides with an image for backgrownd, a voice over and a "Click Button" to be used like a Simulation.  If the "Click Button" is not clicked on, I want the slide to act like a Quiz and be marked as wrong.
    At the end of the Quiz I want the Pass/Fail function to be: If Pass: "Continue" then if Fail "Jump to Slide". This Failed Slide would jump to the last slide; the end of the project with a "Thank you".  If Passed, the project continues to display more information on other slides.
    The tests I have done the Pass/Fail seems to not be working.  If failed, the project continues just like the Passed even though the "Pass/Fail" options are marked correctly.
    What do you think?

    Also check the options under Quiz > Settings > Required.  It could be that you have it set to something other than Optional which might be restricting things.

  • Tracking Pass/Fail on LMS

    I am trying to track pass/fail on a Captivate project and I
    am having difficulties with the LMS. It is tracking as completed
    after just the first slide, before even taking the quiz. I have my
    quiz settings set to report "Quiz results only", Pass/Fail, Report
    to LMS as Percent, Reporting level is "Report Score", the Pass/Fail
    options are set to 75%.
    NOTE: I have allowed the user 2 attempts to pass. Could this
    be what's giving me trouble?

    Look at the Project > Advanced Interaction panel and see
    if you have
    score values setup throughout your lesson as you should.
    It sounds like perhaps you have a low number of scored
    interactions and
    that 75% is being met as soon as they answer the first
    question
    correctly? Compare the Advanced Interaction results to your
    Quiz
    Preferences > Pass/Fail options and see what shakes out...
    Erik
    BriarPatch1 wrote:
    > I am trying to track pass/fail on a Captivate project
    and I am having
    > difficulties with the LMS. It is tracking as completed
    after just the first
    > slide, before even taking the quiz. I have my quiz
    settings set to report
    > "Quiz results only", Pass/Fail, Report to LMS as
    Percent, Reporting level is
    > "Report Score", the Pass/Fail options are set to 75%.
    >
    > NOTE: I have allowed the user 2 attempts to pass. Could
    this be what's giving
    > me trouble?
    >
    Erik Lord
    http://www.capemedia.net
    Adobe Community Expert - eLearning
    http://www.adobe.com/communities/experts/
    http://www.awaretips.net -
    Authorware Tips!

  • Sending a return value without pass/fail message

    Hi,
    I'm using the COM adapter to talk TestComplete application.
    I have a string value test step to evaluate a return value and determine if it's pass or fail (e.g. return 0 = pass, return 1 = fail).  This works perfectly.
    However, I have added another return value (e.g. return 2), the code executed fine, but it keeps displaying that step as "Failed", which is annoying.  Is there a way to suppress this failed message?
    Basically, here's what I need to do, return 0= display pass, return 1 = display fail, return 2 = display no message or just display done.
    Thanks so much!
    Solved!
    Go to Solution.

    The String Value Test step type only evaluates to Passed or Failed.  It does not do a tristate.  In fact any of the step types in the Tests (Pass/Fail Test, Numeric Limit Test, Multiple Numeric Limit Test, String Value Test) will always be pass or fail.  They never evaluate to just done.
    The Status Expression is what makes a step Pass or Fail.  Unfortunately with the default steps you cannot change the Status Expression.  You have to create a custom step type to do that.
    What are you setting your Step.Result.String to?  Or did you change the Data Source?
    Here are 3 options:
    1- Create a custom step based on the String Value Limit test and change the Status Expression such that you get your desired behavior
    2- Use and Action step and change the Status Expression such that you get your desired behavior.
    3- Continue using the step you are using but in the Post-Expression place some logic which turns off the Step.StepFailCausesSequenceFail if you are in return 2 state
    http://zone.ni.com/devzone/cda/tut/p/id/4302
    Hope this helps
    jigg
    CTA, CLA
    teststandhelp.com
    ~Will work for kudos and/or BBQ~

  • Count the Pass/ Fail in a DAQ reading testing ?

    I want to write the while loop to read the DAQ signal. Each time, the DAQ input signal is compared with the spec. limit. The signal that is within the limit will be recorded as " Pass " while the signal out of spec. will be recorded as " Fail ". At the end of the testing ( I break the program), the testing result will be displayed on the testing panel.
    I believe my question is similar to the programming of " A=A+1 ". Can anybody help me to resolve the problem ?

    Tian,
    Are you looking for just the number of passed and number failed measurements or a record of each individual test. If you are only concerned with the number of passed/failed tests then use a shift register. Right-click on the border of the loop and choose Add Shift Register. Then use a case structure. If it passed the test (true case) then increment the value. If it did not pass the test (false case) then just wire the value straight through the case structure. To get the number that failed simply subtract the number that passed from the iteration count.

  • How can assign a PASS/FAIL test to each element of a cluster?

    I would like to pass a cluster from Labview 8.6  to TestStand 4.1.1 and apply a pass/fail test to each boolean element of the cluster.  For elements of the cluster that are string values, I would like to apply a string value test.  For elements of the cluster that are numeric values, I would like to apply numeric value or limit tests.  How can I do this? 

    Hi epsilon-delta,
    The best way to do this would be to create a custom data type from that cluster and store the output of your vi to a local variable. You can then add Pass/Fail Test, Numeric Limit Test and String Value Test steps (using the None adapter) after the Action step to test each element of the cluster.
    Eric B.
    National Instruments

  • Can the sequence pass/fail status be controlled in Test Stand?

    I have a sequence that contains a DoWhile loop. Within the loop are 5 or 6 steps that are executed. If the result of those steps is a fail the loop runs again until it passes or meets its allowable number of iterations. My problem is that if an iteration fails and then passes the result for the sequence is failed. Can the pass/fail status be controlled via an expression etc? I know I can create a subsequence and loop in it with the properties. I am just curious to know if there is a way it can be done like I have explained.
    Thanks in advance for any help,
    Troy

    Troy,
    You might be able to use the RunState.ThisContext.SequenceFailed property.  Set this equal to False after your loop runs, and the sequence should report as if it passed.
    John B.
    Applications Engineer
    National Instruments

  • I want to display the number of Passed / failed tests in Testtand xml report , please help me

    Hi, I want to display the number of Passed / failed tests in Testtand xml report , please help me.
    My requirement is later i should be able to display these number of passed/failed test in report header too.
    Solved!
    Go to Solution.

    I have attached a sequence to give you an idea of how to get the Result count (I didn't check it's working ), I have used "SequenceFilePostResultListEntry" callback to calculate the Result Count,
    and the Result Count is passed to the "ModifyReportHeader" through File Globals,  where you can edit the XML file to show it in the Repot (follow the link in my previous post).
    Hope this helps
    AshwiN,
    Attachments:
    Get Result Count.seq ‏8 KB

  • Can I build a test over several pages, and give instant pass/fail feedback.

    I want to build a test over several pages.  I would like to have buttons where you click on the answer, and at the end of the test show pass/fail and answers to failed questions.  I also want to show the percentage to the test answers, such as 76% out of 100%.  I would like to adjust the percentages depending on the  amount of answers per questions.  Let's say there are 7 questions for the test, a passing score would be 71%, if there are 5 questions, a passing score would be 80% and so forth.
    I am sure there are many questions regarding this, feel free to ask any and I will answer as much as possible.
    I have Keynote 09, Numbers 09 and Pages 09, I would prefer to work with Keynote and hopefully use  the other  two as needed.

    chrisfrommarion wrote:
    Thanks, while it worked on yours it doesn't work on mine, any other ideas?
    Are you holding down the "d" key, all by itself, just after the system shuts down for a restart?
    And how long did you hold it down?
    I held mine down for about 40-50 seconds. Then I lookat at my routers WAN light and it was flickering so I let up on the d key.
    If that does not work then maybe there is something wrong with your MBP. Is that the reason you want to run this test, something isn't right with your unit?

  • In LV, using TS API, how do I specify the VI module for a Pass/Fail step?

    I am writing a LabView VI which will let a user write a TestStand sequence. Each step in the seq is the same type, a Pass/Fail test. I need to specify for each step the LabView VI module location, and an input buffer. I am having difficulty finding out how to specify the module for the step using the TestStand API in LabView. How to do this????

    Hi Drew,
    I think Dennis might have come at this from the wrong angle.
    Have a look at the example that ships with TestStand under
    \Examples\SequenceBuilderTool\LabVIEW
    This might give some clues, but I'd suggest posting this question over to the TestStand discussion Forum as suggested by Dennis.
    There'll be a few more people there who can help a bit more in-depth
    My personal suggestion is to create a custom step type that calls the .vi's and then you can create the steps in a new sequence file using your tool easily. This could be cumbersome for you if you already have a lot of steps setup.
    Better then to create a resource to get the steps out of to clone them by creating a sequence file which is full of your steps. When you have a handl
    e to a cloned step, you can change properties about it in memory, before putting it into the sequence (i.e. change it's name and stuff)
    Otherwise you need to get the type and create a new step of this type.
    If all you're stugling with is the module to call, then for the step you have a handle to, change the
    you can either change the "Step.TS.SData.ViPath" as a string (TestStand set property value String with a reference to the sequence_context) or if you wanted to bring up the specify module dialog, if you have a handle to the step itself, call the method SpecifyModule.
    Hope that helps
    Sacha.
    // it takes almost no time to rate an answer

  • Need help with pass/fail expression​.

    I've got an array of strings parameter called Parameters.CAN_Switch_Channel_Name[0] (equal to a string) which gets passed into a subsequence.  During run time when this subsequence gets called, a new parameter gets created which gets named the string which is passed in through Parameters.CAN_Switch_Channel_Name[0].  So for instance if Parameters.CAN_Switch_Channel_Name[0] is equal to CC_VS_Spd.CC_Set_Switch, then a parameter will get created at run time called: Parameters.CC_VS_Spd.CC_Set_Switch.  I want to create a pass/fail expression to check the new parameter.  I am not sure how to build the expression.  I tried:
    Parameters."Parameters.CAN_Switch_Channel_Name[0]"​  This didn't work!
    Parameters.(Parameters.CAN_Switch_Channel_Name[0])​  This didn't work either!
    I need a way of pulling out the value in Parameters.CAN_Switch_Channel_Name[0] and appending it to Parameters. to reference the parameter created at run time.
    Any ideas?

    Hi,
    I read this a few times now, and just when I think I got ......
    In your SubSequence you have a parameter which is called CAN_Switch_Channel_Name[] which is an array of strings.
    Now do you wish to change the Name of this string array from CAN_Switch_Channel_Name to CC_VS_Spd.CC_Set_Switch or is it that the contains of the array is set to eg
    CAN_Switch_Channel_Name[0] = "CAN_Switch_Channel_Name to CC_VS_Spd.CC_Set_Switch"
    and is this done in the Sequence or before the sequence ist called.
    Maybe a small example of what you are trying to achive, even if its not working may help.
    Regards
    Ray Farmer
    Regards
    Ray Farmer

Maybe you are looking for

  • Apple display with Macbook Pro

    I'm thinking about getting an Apple display for use with a Macbook pro. I'm looking for any observations from those of you using the same configuration. Where do you put your Macbook while using the monitor? Can you sit it on top of the base foot of

  • Rendering Issue with HD

    Hey everyone, I have a problem. I've been cutting DVCPRO HD 720p60 (Panasonic HVX 200 shot at 24) in the newest Final Cut and whenever I have to render a clip in my timeline, the clip content shifts backwards one frame throwing the picture out of syn

  • Resizing window below the Dock

    I know this is weird, but I swear I've done it on another machine. I'm using Leopard 10.5 and for this example I'll choose Calendar. This problem happens to most windows. When I maximize calender, either but hitting the green button, or dragging the

  • Dynamic Title Tags in Webcenter

    This is a newbie question. But I am going to be creating a complete external commercial website using Webcenter/jDev. I see nothing in the documentation about search engine optimization, and more importantly, creating dynamic URL-based title tags. Wi

  • Future Dated Invoices and Payments

    Hello Can you give me some inputs about Future Dated Invoices and Payments..? What its importance? What would the effect on the AP Trial Balance and Account Analysis if I use it? Thanks Grace