AirPlay certification: Unable to execute test cases 3.2.1.3 and 3.2.1.5

Hi everyone,
I'm following the test cases specified in the AirPlay Product Compliance Test R6 Document for AirPlay. I'm unable to execute the below test cases when I'm testing for airPlay certification. There is no option in the latest Airport Utility on Macbook for executing these test cases,
Is it mandatory to execute and pass these test cases? How do I do it?
3.2.1.3 Link Layer Connection 10 Mbps/Full Duplex
3.2.1.5 Link Layer Connection 100 Mbps/Full Duplex
Thanks in advance for your help.
Sripad

Diane,
Thank you very much for the response.
I will email you the screen shots since I can not attach them to this post.
You will see in the screen shots that I don't see either a green or red light after I have clicked on the record topic icon from UPK Developer. Even after I put focus on the title bar of the app I am recording and press the printscreen button, there is no green or red light.
But now, when I do press the printscreen button, UPK does actual record something. That is further than I have gotten before.
Before I posted the forum, I was not able to record anything after several attempts and also several re-installations. But I was told that if I don't see the green or red light, I won't be able to record anything.
I am now wondering if the use of Hypersnap 6 could be impacting UPK? I use Hypersnap a lot and when I pressed the printscreen button, Hypersnap also takes a screenshot.
My question now is whether the lack of green and red lights will impact my use of UPK?
Thanks
Chris
Edited by: user455985 on Dec 22, 2010 2:04 PM

Similar Messages

  • Unable to execute Test-Mailflow cmdlet. Could not find any available Domain Controller

    I started receiving this error in SCOM while testing Exchange 2007 Mail flow. Normally the tests work just fine, and when I execute"test-mailflow" on the offending server, it works just fine. The alert doesn't have enough details to find out what
    is wrong. Can anyone help out?
    Alert Description:
    Unable to execute Test-Mailflow cmdlet. Could not find any available Domain Controller
    Environment: SCOM 2012 R2 UR4
    MP: Exchange 2007 MP, latest update
    Monitor: Exchange 2007 Test Local Mail Flow Monitor
    Brody Kilpatrick

    Hi,
    After you execute"test-mailflow" on the offending server, please also look into event logs to check is there any errors or warnings. SCOM is a tool to collect information, from my point of view, this message is collected from Exchange server, so
    you may need to check event viewer.
    In addition, please also make sure your Exchange environment is under monitored. And have you followed the Exchange 2007 MP guide to configure the MP?
    Make sure the Client Access Server is under monitored, you may check the Application Event log on the Client Access Server, and make sure there are no errors or warnings indicating a problem with the Client Access Server.
    Regards,
    Yan Li
    Please remember to mark the replies as answers if they help and unmark them if they provide no help. If you have feedback for TechNet Subscriber Support, contact [email protected]

  • How can I execute the test case written excel sheet

    Hello
    I have written the test cases  in excel sheet. Please tell any one how to execute test case from labview.please send some sample vi

    I'm sorry, but you've provided zero information on what you have. All you've said is that you have an Excel "sheet" with test cases, and that it's written in VBA. What does that even mean? What's written in VBA? Are you trying to call the VBA code? Are you trying to rewrite the VBA code in LabVIEW? Unless you provide a better explanation of what you have and what you are trying to do, we can just keep playing a guessing game.

  • Unit test cases run twice on TFS2012 build server

    Hi,
    I have enabled the build definition to run the test cases using Visual Studio Test Runner in TFS2012 build server. Once it runs I can see that each test case ran twice. Suppose if I have 5 test cases, 5 will be passed and same 5 will be failed. Why it ran
    for second time and failed? Please advice.
    Below is my activity log. If you notice at the end of the below log, same test cases ran twice. First all are passed and then failed.
    If Not DisableTests00:00:21
    Inputs
    Condition: True
    Run Tests00:00:21
    If Not TestSpecs Is Nothing00:00:21
    Inputs
    Condition: True
    For Each TestSpec in TestSpecs00:00:21
    Inputs
    Values: Run tests in test sources matching **\*test*.dll, Target platform: 'X86'
    Try Run Tests00:00:21
    Handle Test Run Exception00:00:00
    If testException is NOT TestFailureException00:00:00
    Inputs
    Condition: False
    Set TestStatus to Failed00:00:00
    Inputs
    LabelName:
    SourceGetVersion:
    KeepForever: False
    DropLocation:
    Status: None
    LogLocation:
    BuildNumber:
    Quality:
    TestStatus: Failed
    CompilationStatus: Unknown
    If spec.FailBuildOnFailure00:00:00
    Inputs
    Condition: False
    Get Requests Approved for Check In00:00:00
    Outputs
    Result: IQueuedBuild[] Array
    Mark Requests for Retry00:00:00
    Inputs
    Behavior: DoNotBatch
    Requests: IQueuedBuild[] Array
    Force: False
    If spec Is AgileTestPlatformSpec00:00:00
    Inputs
    Condition: True
    Run Visual Studio Test Runner for Test Sources00:00:00
    Assign spec to agileTestPlatformAssembly00:00:00
    Inputs
    Value: Run tests in test sources matching **\*test*.dll, Target platform: 'X86'
    Outputs
    To: Run tests in test sources matching **\*test*.dll, Target platform: 'X86'
    Find Visual Studio Test Platform Test Assemblies00:00:00
    Inputs
    MatchPattern: d:\Builds\8\Test\OfficeLink\src\**\*test*.dll
    Outputs
    Result: System.Linq.OrderedEnumerable`2[System.String,System.String]
    If Visual Studio Test Platform Test Assemblies Found00:00:00
    Inputs
    Condition: True
    If agileTestPlatformAssembly.HasRunSettingsFile00:00:00
    Inputs
    Condition: False
    Run Visual Studio Test Runner for Test Sources00:00:00
    Inputs
    ResultsDirectory:
    TreatTestAdapterErrorsAsWarnings: False
    UpdateFrequencyTimeout: 30
    RunSettings:
    Platform: Any CPU
    TestSources: System.Linq.OrderedEnumerable`2[System.String,System.String]
    KeepAlive: False
    UpdateFrequency: 5000
    TestCaseFilter:
    PublishResults: True
    RunName:
    Flavor: Debug
    ExecutionTimeout: 0
    ExecutionPlatform: X86
    DisableAutoFakes: False
    Passed TestMethod1
    Passed TestMethod1
    Passed RegisteredTypeShouldBeRegistered
    Passed RegisteredTypeLifeTimeManagerIsTryShouldBeRegistered
    Passed RegisteredTypeWhithoutNameShouldBeRegistered
    Passed RegisteredInstanseShouldBeRegistered
    Passed RegisteredInstanseWithoutNameShouldBeRegistered
    Passed RegisteredInstanseShouldBeResolved
    Passed AllRegisteredInstanseShouldBeAllResolved
    Failed RegisteredTypeShouldBeRegistered
    Failed RegisteredTypeLifeTimeManagerIsTryShouldBeRegistered
    Failed RegisteredTypeWhithoutNameShouldBeRegistered
    Failed RegisteredInstanseShouldBeRegistered
    Failed RegisteredInstanseWithoutNameShouldBeRegistered
    Failed RegisteredInstanseShouldBeResolved
    Failed AllRegisteredInstanseShouldBeAllResolved
    Test Run Completed. 16 tests executed.
    Test Run Failed.

    Hi DHegde,  
    Thanks for your post.
    You’re using TFS 2012 Update 4 and VS 2012 Update 2?
    Please manually build your project and run test project using VS 2012 on your client, then check the test case result in Test Result window, the test cases be executed as expect? 
    If everything works fine when manually run your test project using VS 2012, which build process template selected in your build definition? If it’s not TFS 2012 default build process template, please try use default build process template in your current
    build definition, then queue build definition and view test result in TFS Build log again.
    If you’re using default TFS 2012 build process template in your current build definition, but still received the same issue, how did you configure the test in your build definition, please share the detailed configure screenshot(s) here.
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Interesting test case

    Hi,
    I played a little on my test machine and get interesting results so if someone could explain me what happens here I would be grateful.
    It is obvious that some data corruption happened but still interesting situation.
    OS: Linux 32bit
    Oracle: 10.2.0.2.0
    TEST CASE:
    1. Created tablespace and one table in that tablespace:
    SQL> create tablespace test_tbs datafile '/oradata/tbs01.dbf' size 75M autoextend on next 10M maxsize 512M;
    Tablespace created.
    SQL> create table objtab tablespace test_tbs as select * from dba_objects where 1=2;
    Table created.
    2. Made two more copies of tbs01.dbf datafile:
    SQL> !cp tbs01.dbf tbs02.dbf
    SQL> !cp tbs01.dbf tbs03.dbf
    3. Insert some rows into table objtab:
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> /
    50417 rows created.
    SQL> /
    50417 rows created.
    SQL> commit;
    Commit complete.
    SQL> select count(*) from objtab;
    COUNT(*)
    151251
    4. Deleted tbs01.dbf:
    SQL> !rm tbs01.dbf
    5. Insert still works:
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> commit;
    Commit complete.
    6. Renamed tbs02.dbf to tbs01.dbf:
    SQL> !mv tbs02.dbf tbs01.dbf
    7. Inserted new rows:
    SQL> select count(*) from objtab;
    COUNT(*)
    302502
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> commit;
    Commit complete.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> commit;
    Commit complete.
    8. Check size of datafile:
    SQL> !du -hs tbs01.dbf
    *96M tbs01.dbf*
    SQL> select count(*) from objtab;
    COUNT(*)
    756255
    9. Deleted datafile tbs01.dbf and renamed tbs03.dbf to tbs01.dbf:
    SQL> !rm tbs01.dbf
    SQL> !mv tbs03.dbf tbs01.dbf
    10. Insert more rows and executed "alter system checkpoint":
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> insert into objtab select * from dba_objects;
    50417 rows created.
    SQL> commit;
    Commit complete.
    SQL> alter system checkpoint;
    System altered.
    SQL> select count(*) from objtab;
    COUNT(*)
    907506
    11. When I check the size of tbs01.dbf it is smaller then before despite of more rows I inserted. How come? Where are all this rows stored?
    SQL> !du -hs tbs01.dbf
    *86M tbs01.dbf*
    12. Now try to offline tablespace and then I get errors in alertlog:
    SQL> alter tablespace test_tbs offline normal;
    alter tablespace test_tbs offline normal
    ERROR at line 1:
    ORA-00603: ORACLE server session terminated by fatal error
    ALERT LOG
    Errors in file /oracle/admin/um/udump/um_ora_501.trc:
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01122: database file 6 failed verification check
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01208: data file is an old version - not accessing current version
    Fri May 29 09:01:31 2009
    ORA-600 signalled during: alter tablespace test_tbs offline normal...
    Fri May 29 09:01:31 2009
    Errors in file /oracle/admin/um/udump/um_ora_501.trc:
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01122: database file 6 failed verification check
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01208: data file is an old version - not accessing current version
    Fri May 29 09:01:37 2009
    Errors in file /oracle/admin/um/udump/um_ora_501.trc:
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01122: database file 6 failed verification check
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01208: data file is an old version - not accessing current version
    Fri May 29 09:01:43 2009
    Errors in file /oracle/admin/um/udump/um_ora_501.trc:
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-00600: internal error code, arguments: [krhpfh_03-1208], [fno =], [6], [fecpc =], [4], [fhcpc =], [3], []
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01122: database file 6 failed verification check
    ORA-01110: data file 6: '/oradata/tbs01.dbf'
    ORA-01208: data file is an old version - not accessing current version
    13. One more check:
    SQL> select tablespace_name, status from dba_tablespaces;
    TABLESPACE_NAME STATUS
    SYSTEM ONLINE
    SYSAUX ONLINE
    USERS ONLINE
    UNDOTBS2 ONLINE
    TMP ONLINE
    TEST_TBS ONLINE
    7 rows selected.
    SQL> select count(*) from objtab;
    COUNT(*)
    907506
    14. Restart database:
    SQL> shutdown immediate;
    Database closed.
    Database dismounted.
    ORACLE instance shut down.
    SQL> startup;
    ORACLE instance started.
    Total System Global Area 1224736768 bytes
    Fixed Size 1267188 bytes
    Variable Size 1006635532 bytes
    Database Buffers 201326592 bytes
    Redo Buffers 15507456 bytes
    Database mounted.
    Database opened.
    15. Try to count(*) from table but it says that object no longer exists:
    SQL> select count(*) from objtab;
    select count(*) from objtab
    ERROR at line 1:
    ORA-08103: object no longer exists
    SQL> select tablespace_name, status from dba_tablespaces;
    TABLESPACE_NAME STATUS
    SYSTEM ONLINE
    SYSAUX ONLINE
    USERS ONLINE
    UNDOTBS2 ONLINE
    TMP ONLINE
    TEST_TBS ONLINE
    7 rows selected.
    16. If object does not exist, how come I get this results (probably leftovers in data dictionary):
    SQL> desc objtab;
    Name Null? Type
    OWNER VARCHAR2(30)
    OBJECT_NAME VARCHAR2(128)
    SUBOBJECT_NAME VARCHAR2(30)
    OBJECT_ID NUMBER
    DATA_OBJECT_ID NUMBER
    OBJECT_TYPE VARCHAR2(19)
    CREATED DATE
    LAST_DDL_TIME DATE
    TIMESTAMP VARCHAR2(19)
    STATUS VARCHAR2(7)
    TEMPORARY VARCHAR2(1)
    GENERATED VARCHAR2(1)
    SECONDARY VARCHAR2(1)
    SQL> create table objtab tablespace test_tbs as select * from dba_objects;
    create table objtab tablespace test_tbs as select * from dba_objects
    ERROR at line 1:
    ORA-00955: name is already used by an existing object
    17. To drop object (can't drop):
    SQL> drop table objtab;
    drop table objtab
    ERROR at line 1:
    ORA-08103: object no longer exists
    18. Clean up:
    SQL> drop tablespace test_tbs including contents and datafiles;
    Tablespace dropped.
    Best Regards,
    Marko

    Hi Uwe,
    as I said before my intention was to understand behavior of Oracle and Linux in scenario like this.
    I thought about what will happen if some Linux admin by mistake moves live datafile to another location and after 5-10 mins, when he realizes his mistake, he moves the datafile to old (original) location.
    Will Oracle notice some errors in alert log? (In my test case I didn't receive any message in alert log)
    How will I know (as a DBA) what was done with this datafile if Linux admin does not say anything to me?
    Is any damage made to datbase ? (Probably)
    etc...
    When I am on my test machine I like to do all kind of stuff and try anything that comes on my mind. It isn't important to me if this scenario has any connection to real world problems. I enjoy doing this so I like to spend some of my time on this test cases.
    Anyway thanks for your comment.
    Regards,
    Marko
    Edited by: msutic on May 29, 2009 1:03 PM

  • Report to run CATT test case

    Hi Experts,
    I have a requirement where i need to create a report program to run the   CATT test case 'XYZ'.
    CATT Test case 'XYZ' mass updates pricing for materials via VK11 for condition type PR00, sales area GB20/IC/01.
    As the business should not have permanent authorization to run CATT they requires a report to execute test case 'XYZ' to be developed.
    I have no idea how to run CATT test case through report program. Any pointers will be highly appreciated.
    Thanks & Regards,
    Swati

    Hi Balu,
    Thanks for your feedback. But i am still not clear.Could you please let me know the procedure in steps.
    I also have FS in which functional mention to upload file data and run the test case for some condition type and sales area.
    File have following fields:
    Material, Rate, Pricing Unit, Valid on, Valid to
    Regards,
    Swati
    Edited by: Swati Garg on Mar 7, 2011 6:32 PM

  • How to store test cases in multiple languages?

    We need to provide test cases in German an English and I would like to avoid having two projects.
    My idea is to use a template project which allows the translation of the process tree, and then have both test cases in one document (maybe using a word makro to jump to the needed language if that works).
    Do you know a better way to manage test cases in different languages?

    We need to provide test cases in German an English and I would like to avoid having two projects.
    My idea is to use a template project which allows the translation of the process tree, and then have both test cases in one document (maybe using a word makro to jump to the needed language if that works).
    Do you know a better way to manage test cases in different languages?

  • Problem at the time of Implementing Text Matching Test Case in OFT

    hi,
    I had add Text Matching Test Case on login of the application for the username and password. and If the Test Case fail on that screen,then it should not allow to go further in the application.
    As currently at the time of Playback, it is allowing to go further and in the Result Report it display the case failed.

    Actually my ques is that suppose at the time of recording i enter the username as abcd and password as 123.
    When the recording is done . I Insert a Text Matching test case for both the username and password where i put the condition for the username that select text should be present as "def" . And the Test Case failed .
    So i want to know that if the test case failed on Login. Whether it should be move forward at the time of playback?

  • Move test case with referenced items to another project

    How can I move a test case with all referenced contexts and verifications from a project to another project? At the moment I have to move each single item. How can I do this more efficient?

    Hi Markus,
    You can Export/Import it in a Test Portable format.
    All contexts will be included.
    Kind regards,
    Ulyana.

  • WCF debug Test-case issues.

    This seems to be a weird issue, I had solution with a WCF Service and Unit Test Project.
    In Unit test, I had some testcases which will step into WCF service. When I start the service, and go into Test-case (Test explorer) all test-cases are disabled, and in Test => all Debug / RunAll tests all options are disabled.  So I am not able
    to debug testcases when WCF Service is running this is the issue. I am using VS2013 Ultimate with Update3 installed.
    I am rightclicking on WCF Service debug start new instance, then go into Test-case (Test-explorer) when I right click on test-case all options are disabled and so is in Test menu. 
    Thanks In Advance,
    KSR

    Hi KSR,
    To debug the WCF Service from the Test project, one idea is that you could creates and opens the service host in test project code.
    A similar thread about this issue, reference:
    http://stackoverflow.com/questions/2790901/debugging-visual-studio-2010-unit-test-and-wcf-service-in-one-ide-instance
    A blog also shared some suggestions here:
    http://blogs.msdn.com/b/ploeh/archive/2006/12/04/integrationtestingwcfservices.aspx
    Best Regards,
    Jack
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • GRC - Test Cases

    Hi All,
    We are approaching SIT phase for GRC. Are there any standard scripts for GRC in specific during testing.
    I assume that below are SIT test cases. But if I am missing something, experts can advise.
    All my MSMP workflows - Using Template based requests (We are using Templates for each request type)
    checking Msmp workflows buttons (Approve, Reject, Forward, Return)
    checking UAR workflows
    Checking Risk analysis during workflows - For few important risks
    checking EAM workflows and how log review report workflows work
    Checking all Email notifications (If they are proper)
    Are there any other scenarios than the mentioned one's which will be crucial during roll out?
    Please suggest
    Thanks,
    Sai.

    Dear Sai,
    depends on your setup. Basically you have to define the test cases based on your requirements and setup. Hence it is difficult to tell you what needs to be tested.
    Best practise from my experience is to define all cases and expected output. Best if you think in processes and define test scripts for each process. In the end the process has to work from A to Z.
    Regards,
    Alessandro

  • Can I re-use a parameterized manual test case with different set of parameters in a different test plan?

    Can I re-use a parameterized manual test case with different set of parameters in a different test plan without impacting the parameters in the first test plan?

    Hi ssteele_1,
    Thank you for posting in MSDN forum.
    Generally, I know that if you want to re-use a parameterized manual test case and without impacting the parameters in the first test plan.
    I suggest you could try to right click this parameterized manual test case ->Select the Create copy and add to suite option like the following screen shot.
    Then the parameterized manual test case can be re-used as a new test case in this MTM and it will did not impact the original parameterized manual test case.
    Best Regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Problem in executing JFCunit test case

    Hi all,
    I have problem in executing a simple test case that checks for input blank fields entered.
    Below is the code for the test class that i am trying to run, but i am unable to execute it. Please can anyone try to resolve this.
    package unittest.com.erp;
    import junit.extensions.TestSetup;
    import junit.extensions.jfcunit.*;
    import junit.extensions.jfcunit.finder.*;
    import junit.extensions.jfcunit.eventdata.*;
    import junit.framework.Test;
    import junit.framework.TestSuite;
    import junit.textui.TestRunner;
    import com.erp.client.swing.ClientLoginDialog;
    import com.erp.client.swing.workspace.ClientWorkspaceFrame;
    import com.erp.client.swing.workspace.data.LoginResults;
    import javax.swing.*;
    public class LoginScreenTest
        extends JFCTestCase {
      private ClientLoginDialog loginScreen = null;
      private JButton login = null;
      private JButton cancel = null;
      private JTextField username = null;
      private JTextField password = null;
      private JFCTestHelper helper = null;
      LoginResults loginSuccess = null;
      ClientWorkspaceFrame workspaceFrame = null;
      public LoginScreenTest() {
      public LoginScreenTest(String name) {
        super(name);
      public static Test suite() {
        return new StartApp(new TestSuite(LoginScreenTest.class));
      private static class StartApp
          extends TestSetup {
         * Construct the test decorator, which starts the application     *
         * @param test
         *          Test case.
        public StartApp(final Test test) {
          super(test);
         * Start the LoginScreenTestapplication.
        public void setUp() {
          new Thread(new Runnable() {
            public void run() {
              try {
                new LoginScreenTest().setUp();
              } catch (Exception e) {
          }).start();
          try {
            Thread.currentThread().sleep(10000);
          } catch (InterruptedException ex) {
         * Tear down the LoginScreenTest application.
        public void tearDown() {
      protected void setUp()
          throws Exception {
        super.setUp(); // Choose the text Helper
        setHelper(new JFCTestHelper()); // Uses the AWT Event Queue.
        // setHelper( new RobotTestHelper( ) ); // Uses the OS Event Queue.
        ClientWorkspaceFrame workspaceFrame = ClientWorkspaceFrame.getHandle();
        // loginScreen = new ClientLoginDialog( "LoginScreenTest: " + getName());
        loginSuccess = ClientLoginDialog.login();
        loginScreen = ClientLoginDialog.login(new JFrame(), true);
        if (loginSuccess != null && loginSuccess.isSuccess()) {
          workspaceFrame.initFrame();
          workspaceFrame.loggedInUser = loginSuccess.getName();
          workspaceFrame.loggedInFullUserName = loginSuccess.getFullLoginUserName();
          workspaceFrame.loggedInUserPassword = loginSuccess.getPassword();
          workspaceFrame.setVisible(true);
          workspaceFrame.validate();
            flushAWT();
      protected void tearDown()
          throws Exception {
        login = null;
        cancel = null;
        username = null;
        password = null;
        loginScreen = null;
        flushAWT();
        // getHelper.cleanUp( this );
        super.tearDown();
      public void testInitialState() {
        NamedComponentFinder finder = new NamedComponentFinder(JButton.class, "Cancel");
        finder.setName("Cancel");
        JButton exitButton = (JButton) finder.find(loginScreen, 0);
        assertNotNull("Could not find the Exit button", exitButton);
        NamedComponentFinder finder1 = new NamedComponentFinder(JButton.class, "Login");
        finder1.setName("Login");
        JButton enterButton = (JButton) finder1.find(loginScreen, 0);
        assertNotNull("Could not find the Enter button", enterButton);
        getHelper().enterClickAndLeave(new MouseEventData(this, enterButton));
        assertEquals("", workspaceFrame.loggedInUser);
        assertEquals("", workspaceFrame.loggedInUserPassword);
      public static void main(final String[] args) {
        TestRunner.run(LoginScreenTest.suite());
    Thanks & Regards,
    VishalMessage was edited by:
    vishal_vj
    Message was edited by:
    vishal_vj
    Message was edited by:
    vishal_vj
    Message was edited by:
    vishal_vj

    hi All,
    can any one guide me how to run the test casese in JFCunit ? as when i tried to run thru command prompt its not recognising the main class at all.
    Now i m trying with eclipse problem here is that it is showing error to this line of code
    DialogFinder dFinder = new DialogFinder(loginScreen);
    error is constructor is undefined?
    So looking for some solution.
    with regards
    kin

  • Where should i found the executed automation test cases result in TestResult folder?

    we are using VS2013, i have automated some coded ui test cases and executed from test explorer. once completed i am able to view the test result in the Test explorer window. But I am unable to view in the 'TestResult' folder under project location. (In The
    Test result folder created IN & Out sub folder )  I think it should be create .trx file under TestResult folder.  is there any configuration required.
    Please guide me.
    Thanks in Advance. 

    Hi Divakar,
    Thank you for posting in MSDN forum.
    As far as I know that when we automated coded ui test cases from VS IDE, it is default that there have no the .trx file be created in the TestResult folder. However, there have a replace workaround that we can automate coded ui test cases using MSTest.exe
    in command line. And then we can get the .trx file be created in the TestResult folder.
    So you can execute the /resultsfile command to get the .trx file using mstest.exe in command line.
    Reference:
    https://msdn.microsoft.com/en-us/library/ms182489.aspx?f=255&MSPPError=-2147217396
    Best Regards,
    We are trying to better understand customer views on social support experience, so your participation in this interview project would be greatly appreciated if you have time. Thanks for helping make community forums a great place.
    Click
    HERE to participate the survey.

  • Executing Junit test case in EM console

    Hi ,
    When i execute Junit test case in EM console, and against the tab test runs my assertion details gets failed and data in both column, expected value and actual value remains the same but in the expected value column the name space gets automitically populated. Am not able to include the namespace in Junit test case as it gets removed at the run time .
    Kindly suggest.

    Were you able to solve this issue ?

Maybe you are looking for