Unit test of FPM applications?

Is there any way to implement/execute some kind of unit tests for floorplan manager applications? I would like to write some automatically reproducible tests but cannot find any solution/framework for it.
Thomas

Dear Thomas,
Nguyen is right. Either you mock the behavior directly in the UI (then no ABAP Unit is needed). Or you are using GUIBBs with feeder classes and use ABAP Unit which requires you to call methods like FLUSH or PROCESS_EVENT in the ABAP unit test manually.
I tried it myself and it works pretty straight forward. All you would have to to is to
resolve all dependencies like the instance of IF_FPM in the INITIALIZE method.
A new feeder's subclass will be subject to the unit test.
Override INITIALIZE there and replace all dependencies by mocked instances
call the UIBB's methods like the FPM would do it: INITIALIZE, FLUSH, PROCESS_EVENT etc.
when needed, also mock parameter input like events (CL_FPM_EVENT)
carry out verifications against the mocked dependencies (e.g. if various methods like RAISE_EVENT or OPEN_DIALOG_BOX in IF_FPM have been called or not, or of certain wire objects have been passed)
Regards,
Uwe

Similar Messages

  • How can I unit test my Swing application?

    Hi All,
    In my application I came across a situation where I have to automate the unit testing of my application.
    Using my application user can able to draw polylines and rectangles on a BufferedImage added to a JFrame.
    Can any one please help/suggest me, how to automate the drawing process (this process consists of drawing the polylines with mouse and few action/mouse events)?
    your help will be highly helpful to me.
    Satya

    What exactly will that unit test prove? That you can mock the drawing of lines?
    Unit test are supposed to test certain parts of your application code - I'd concentrate on unit tests that are actually useful and for example in the future can help to catch regressions.
    If anywhere you can get help creating unit tests of this magnitude, I'd look in the unit testing support for Swing applications. Google for "java swing unit test" to see what you can come up with. Probably you'll need to create some plumbing yourself to make custom painting unit-testable; perhaps you could do something with painting to a BufferedImage in stead of to a Swing component for example.

  • Unit test for J2EE application

    I am writting a Unit test for One J2EE application. The Application is built in such a way that makes unit testing extremely difficult.
    There are 2 things that contribute to the mess.
    1. Sping integration means all the config files are specified in web.xml independently, even though their beans rely on each other across files. End result is in a unit, I cannot try to load a bean because some of its dependencies are missing (ie. they are in a config file that the first file does not include). For this I tried to use AbstractDependencyInjectionSpringContextTests class to set the Spring Application Context when you are not in the flow but didn't success ed.If any one has use this please post the example.
    2. The application is using Errors interface of package org.springframework.validation. To write a test for any validator class you have to pass the object of Error in the validate method with the command object. Now my question is how you can set this Error object when you are not in the flow. For this I tried to use Mock object e.g Errors error;
    Mockery context = new Mockery();
    final Errors errorMock =context.mock(Errors.class);
    //call the validate object with mock object
    classObject.doValidate(cmdObject, errorMock)
    This thing doesn't work. It gives me below error message.
    unexpected invocation: errors.pushNestedPath("")
    no expectations specified: did you...
    - forget to start an expectation with a cardinality clause?
    - call a mocked method to specify the parameter of an expectation?
    Is there any way to get around these hiccups programmatically in unit tests?
    thanks...

    If you are doing unit testing, try to use straight JUnit4 without involving the Spring framework. Given that you do unit testing, you might not need Spring configuration in your unit test at all. You can programmatically instantiate the instance of the class under testing and either programmatically instantiate collaborating objects, or create mock objects for that purpose. If you are doing functional testing, you might need a Spring context after all. Understand that your tests are running in the different context than the complete application, so you would have to create separate application context for your test(s). You might have to go through the existing Spring configuration modules that you created for your application and re-jiggle them a bit so that they can be included both in your application context and your unit test context.
    Hope this helps.

  • Unit testing PL/SQL used in APEX application

    question from my customer :
    I am developing an application in Oracle Application Express and am working on unit tests for the PL/SQL stored procedures and packages that are stored in the underlying database and that are used by the APEX application. These unit tests should run within the SQL Developer Unit Test framework.
    The problem is that some of the PL/SQL code stored in the database uses functions like NV('APPLICATION_ITEM') to access items in the apex application. These do not return any values when I try to execute the PL/SQL within the unit test framework, ie through the backend. While it is good that the NV function does not error, NULL do not really work well in my scenario either (for example when the result of this functions is inserted into a NOT NULL column of a table). I can think of a few workarounds, such as creating my own NV function inside the test schema to return desirable values, but nothing seems a really satisfactory solution. I just wonder if there is any best practice recommendation from Oracle for this scenario - how can I run code that uses APEX-specific functions through the back end. I could not find anything in the APEX documentation for this but I'd be interesting to know if there is any recommendation how to best deal with this case.
    I am using SQL Developer version 4.0.0.13.80

    User[[:digit:]*
    Your PL/SQL Package APIs are poorly designed.
    You need to take Tom Kyte's quote to heart:
    "Application come and application go, but data remains forever"
    In short, you need to separate your database processing code (the stuff you need to unit test) from front-end/middle tier code.
    (repetitiveness is for effect.. not rudeness.)
    As such, The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in APEX.
    The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in .NET.
    The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in JSP.
    The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in Jive.
    The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in Ruby.
    The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in Perl::CGI.
    The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in P9.
    The PL/SQL code that you need to 'UNIT TEST' must work without needing to run in <place latest and greatest thing here>.
    Again, I don't mean to sound rude.  I'm just trying to reinforce the idea that you need to separate database code from middle-tier/front-end stuff.
    Basically, you will need to separate all of your packages into multiple parts.
    a _CORE package (that will be unit tested) that does all the hard work
    an _APEX package for APEX infrastructure (this works with NV()/V(), etc.)
    a _NET package for .NET infrastructure when you need it
    a _JSP package for the JSP infrastructure when you need it
    a _JIVE package for the JIVE infrastructure when you need it
    a _<place latest and greatest thing here> for the <place latest and greatest thing here> when you need it.
    MK

  • Cairgorm 3 sample applications - errors in unit tests

    After the last update the sample applications started to work - they now even show images.
    But the unit tests for Commands are reporting errors on my machine:
    TypeError: Error #1034: Type Coercion failed: cannot convert []@56443a9 to XMLList.
    at asmock.integration.flexunit::ASMockClassRunner/withBeforeClasses
    at org.flexunit.runners::ParentRunner/classBlock(ParentRunner.as:221)
    etc...
    As I am just starting to explore unit testing, I do not know how to debug it.
    Can someone help me?
    Thanks, Zdenek M

    Hi,
    Could you confirm that you are running the unitTests using the InSyncBasicTestRunner.mxml? Also make sure that you don't have another version of FlexUnit4 on your classpath (Project->Properties->Flex Build Path) as the one shipped with the latest version of Flash Builder may conflict with the version of ASMock we are using at this time.
    Let me know if you are still experiencing issues with the test suite.
    Best,
    Nicolas

  • Flash Builder + unit tests = Unknown error generating output application.xml

    Please see my initial forum post here: http://forums.adobe.com/message/3730193
    This forum does not let me mark my initial post as "not answered" (re-opened) and it seems the dev is not going to respond to it as-is (maybe because he does not know)... so I'll re-open it this way
    An update to my initial post:
    I am now seeing this on Flex 4.1 projects as well. Both AIR and Flex Library projects.
    It appears to be some kind of race condition. Sometimes unit tests will launch on the first try, other times it takes 4-5 tries. If I keep the unit tests application XML open in the background, even on successful launches I see that the <content> node is re-written to contain the square brackets.

    Thanks Sudhir. I had just assumed that you were not getting the notifications since the ticket was marked as "answered".
    I appreciate your help in this matter.

  • When building an application unit tests are added to a support directory

    Why are unit tests (.lvtest files) added to a support directory when building an application? I cannot find a way to exclude them from the support directory.

    I'd imagine this is happening to you because they are in the "Dependencies" section of your project file (*.lvproj). Is this correct?
    If so, you should add them to your project file somewhere. Then you will be able to prevent them from being installed. In my experience LabVIEW grabs files it thinks are nesecary from the Dependencies list and installs them in a support directory. 

  • [svn:osmf:] 17456: Restore unit test application to (manual) runnable state .

    Revision: 17456
    Revision: 17456
    Author:   [email protected]
    Date:     2010-08-23 16:51:56 -0700 (Mon, 23 Aug 2010)
    Log Message:
    Restore unit test application to (manual) runnable state.
    Modified Paths:
        osmf/trunk/framework/OSMFTest/OSMFTest.mxml

  • Unit Testing

    Hi All,
    I would like to inform to all, can any one explain to me the concept of Unit testing in FICO and how it should be drawn? If possible send any dcoumentation for this focus to my mail id: [email protected]
    Thanks with regards,
    Bala

    Hi
    refer below reward if helps
    Unit testing is done in bit and pieces. Like e.g. in SD standard order cycle; we do have 1-create order, then 2-delivery, then 3-transfer order, then 4-PGI and then 5-Invoice.  So we will be testing 1,2,3,4 and 5 seperately alone one by one using test cases and test data. We will not be looking and checking/testing any integration between order and delivery; delivery and TO; TO and PGI and then invoice.
    Whrereas System testing you will be testing the full cycle with it's integration, and you will be testing using test cases which give a full cyclic test from order to invoice.
    Security testing you will be testing different roles and functionalities and will check and signoff.
    Performance testing is refered to as how much time / second will take to perform some actions, like e.g. PGI.  If BPP defination says 5 seconds for PGI then it should be 5 and not 6 second.  Usually it is done using software.
    Regression testing is reffered to a test which verfies that some new configuration doesnot adversly impact existing functionality.  This will be done on each phase of testing.
    User Acceptance Testing:  Refers to Customer testing. The UAT will be performed through the execution of predefined business scenarios, which combine various business processes. The user test model is comprised of a sub-set of system integration test cases.
    We use different software during testing. Most commonly use are
    Test Director:  which is used to record requirement, preparing test plan and then recording the progress.  We will be incorporating defects that are coming during these testings using different test cases.
    Mercury Load Runner:  is used for performance testing.  This is an automatic tool.
    What does the following terms means :
    - Technical Unit Testing
    - Functional Unit Testing
    - IntegrationTesting
    - Volume Testing
    - Parallel Testing?
    Technical Unit Testing= Test of some technical development such as a user exit, custom program, or interface. the test usually consists of a test data set that is processed according to the new program.  A successful test only proves the developed code works and that it performed the process as as designed.
    Functional Unit Testing= Test of configuration, system settings or a custom development (it may follow the technical unit testing) These usually use actual data or data that is masked but essentially the same as a real data set. A successful test shows that the development or configuration works as designed and the data is accurate as a result.
    IntegrationTesting= Testing a process, development or configuration within the context of any other functions that the process, development or functionality will touch or integrate . The test should examine all data involved across all modules and any data indirectly affected. A successful test indicates that the processes work as designed and integrate with other functions without causing any problems in any integrated areas.
    Volume Testing= testing a full data set that is either actual or masked to insure that the entire volume does cause system problems such as network transmission problems, system resources issues, or any systemic problem, A successful test indicates that the processes will not slow or crash the system due to a full data set being utilized.
    Parallel Testing= Testing the new system or processes with a complete data set while running the same processes in the legacy system. A successful test will show identical results when both the legacy system and new system results are compared.
    I would also note that when a new implementation is being done you will want to conduct at least one cut over test from the old system to the new and you should probably do several.
    What kind of testings that are carried out in testing server?
    1. Individual Testing ( Individually which we've created)
    2. Regressive Testing ( Entire Process)
    3. Integration Testing ( Along with other integrated modules)
    The 3 types of testing is as follows:-
    1. Unit testing (where an individual process relevant to a SD or MM etc is tested)
    2. Integration testing (where a process is tested that cuts across all areas of SAP).
    3. Stress testing (where lots of transactions are run to see if the system can handle the data)
    http://www50.sap.com/businessmaps/6D1236712F84462F941FDE131A66126C.htm
    Unit test issues
    The unit tools test all SAP development work that handles business object processing for the connector. Also, the unit test tools enable you to test the interaction of your work with the ABAP components of the connector. The test tools allow you to test your development work as an online user (real-time) only.
    Note:
    It is important to understand the differences between testing the connector as an online user and testing the connector as if operating as a background user.
    The differences between testing the connector as an online user or as a background user are described as follows:
    Memory--When testing a business object, the connector must log into the SAP application.
          The connector runs as a background user, so it processes in a single memory space that is never implicitly refreshed until the connector is stopped and then restarted (therefore it is critical in business object development to clear memory after processing is complete). Since you are an online user, memory is typically refreshed after each transaction you execute.
          For more information, see Developing business objects for the ABAP Extension Module. Any problems that may occur because of this (for example, return codes never being initialized) are not detected using the test tool; only testing with the connector will reveal these issues.
    Screen flow behavior--Screen flow behavior is relevant only when using the Call Transaction API. The precise screen and sequence of screens that a user interacts with is usually determined at runtime by the transaction's code. For example, if a user chooses to extend a material master record to include a sales view by checking the Sales view check box, SAP queries the user for the specific Sales Organization information by presenting an additional input field. In this way, the transaction source code at runtime determines the specific screen and its requirements based on the data input by the user. While the test tool does handle this type of test scenario, there is a related scenario that the test tool cannot handle.
          SAP's transaction code may present different screens to online users versus background users (usually for usability versus performance). The test tool only operates as an online user. The connector only operates as a background user. Despite this difference, unit testing should get through most of the testing situations.

  • Unit testing or integration testing?

    Here it says that this example is more of an explanation of framework than best practice because of the conflation of integration/unit testing concepts: http://docs.flexunit.org/index.php?title=Using_Asynchronous_Startup
    My question is, how do you define a line here between unit testing and integration testing? What kind of UI components asynchronous tests are then considered to be unit tests and not yet integration tests? Many books recommend having 100% code coverage with unit tests. How would this example look like, to be considered only a unit test and not integration test?

    The question about which levels of testing to do are really dependent upon you. Let me explain a bit more.
    If you are building an application, and only an application, then functional testing probably covers most of your needs. In an application environment, you have a relatively narrow pathway of options and interactions that a user might follow. So long as you test those options well, your app will probably work as well.
    Where I see the real value of unit testing is testing for the unknowns. If you have a class that will be used in multiple applications in multiple ways, or perhaps it is intended to be a reusable class where others in your organization will use it in the future, then you want lower level testing. You want to have the best coverage you can because you want to know that, regardless of the way someone chooses to use the code, it will likely function or at least fail gracefully.
    The same is true for integration. If I am creating a Flex component, then I can unit test part of it, but not all. So, I want to use integration tests to ensure that it works in as many circumstance as possible. Further, Flex components have a life cycle, so testing them when they are not on the display list isn't really a valid test..
    So, to summarize, I see unit and integration w/Flex about testing possibilities and trying to achieve realistic code coverage for known and possible use cases. I see functional testing as testing those components in very specific use cases.
    Which do you need? Depends. Are you developing reusable components across (n) projects, likely the first 2 and maybe the last. Are you an application developer testing that an application works in documented business cases? Just functional will likely work.
    Hope that helps,
    Mike

  • Unit testing in J2EE environment

    Hi All:
    We have been trying to use Junit for creating unit test scripts, and have been bit successful in unit testing DAOs and Value objects - but problem is testing of components like EJBs and Servlets or even classes like Actions or Commands. Since these components run under application server environment, I am not sure how to unit test them without either deploying them on the actual server or simulating app server and rest of the system.
    I was wondering if people could share their experience in writing unit test scripts especially for J2EE components - like Servlets, JSPs and EJBs. On the similar note, is there any similar tool or API for creating integration test scripts?
    Thanks,

    Well - we already have couple of other servers but that's for beta, QA and Integration testing. Problem is with the unit testing. In unit testing, some piece of code is tested by itself. And I am not sure how I can unit test some of the classes like Servlets, EJBs, JSPs or even dependent classes like Commands or Actions. For example, consider following Action class from Java Pet Store:
    public final class CartHTMLAction extends HTMLActionSupport {
    public Event perform(HttpServletRequest request)
    throws HTMLActionException {
    // Extract attributes we will need
    String actionType= (String)request.getParameter("action");
    HttpSession session = request.getSession();
    // get the shopping cart helper
    CartEvent event = null;
    if (actionType == null) return null;
    if (actionType.equals("purchase")) {
    String itemId = request.getParameter("itemId");
    event = new CartEvent(CartEvent.ADD_ITEM, itemId);
    else if (actionType.equals("remove")) {
    String itemId = request.getParameter("itemId");
    event = new CartEvent(CartEvent.DELETE_ITEM, itemId);
    else if (actionType.equals("update")) {
    Map quantities = new HashMap();
    Map parameters = request.getParameterMap();
    for (Iterator it = parameters.keySet().iterator();
    it.hasNext(); ) {
    String name = (String) it.next();
    String value = ((String[]) parameters.get(name))[0];
    final String ITEM_QTY = "itemQuantity_";
    if (name.startsWith(ITEM_QTY)) {
    String itemID = name.substring(ITEM_QTY.length());
    Integer quantity = null;
    try {
    quantity = new Integer(value);
    catch (NumberFormatException nfe) {
    quantity = new Integer(0);
    quantities.put(itemID, quantity);
    event = CartEvent.createUpdateItemEvent(quantities);
    return event;
    In order to unit test above class from say, JUnit test script class, I will have to pass HttpServletRequest object to its method - I will also have to store corresponding command params and store them in the request object - how will you write unit test script for the above class?

  • Unit Testing and APEX Global Variables

    We've recently started to unit test our database level PL/SQL business logic.
    As such we have a need to be able to simulate or provide output from PL/SQL APEX components in order to facilitate testing of these components.
    Some of the most obvious portions that need simulation are:
    1. The existence of a session
    2. The current application ID
    3. The current page ID.
    We currently handle requirement #1 by using apex_040100.wwv_flow_session.create_new
    We handle 2 and 3 using the apex_application.g_flow_id and g_flow_step_id global variables.
    I'm just wondering, how safe is it for us to use wwv_flow_session.create_new to simulate the creation of a session at testing time for those things which need a session?
    I've also noticed that there are apex_application.get_application_id and apex_application.get_page_id functions whose output is not tied to the global variables (at least in our current version).
    Is it safe for us to expect that we can set these global variables for use in testing or is apex moving to get_application_id and get_page_id functions away from global variables?
    Will there be corresponding set_application_id and set_page_id functions in the future?
    Sorry for the question bomb. Thanks for any help.

    My first question would be why do you need to establish a session to test your PL/SQL?
    wwv_flow_session is a package internal to APEX, and you should probably leave it be.
    The get_application_id procedure you refer to is in apex_application_install, which is used for scripting installation of applications - not get/set of page ID like you're describing.
    If you're uncomfortable using apex_application.g_flow_id, you can use v('APP_ID') or preferably pass the app_id/page_id as parameters to your procedures.
    Your question seems to have a few unknowns, so that's the best I can describe.
    Scott

  • Unit testing and integration testing

    hello 2 all,
                    what is the diff bet unit and integration testing? in sap what is unit teesting consists of and integration testing consists of what?
    is this the work  of test engineers r whose work is this?
    take care
    love ur parents

    Hi Sameer,
    Unit Testing
    A unit test is a procedure used to validate that a particular module of source code is working properly from each modification to the next. The procedure is to write test cases for all functions and methods so that whenever a change causes a regression, it can be quickly identified and fixed. Ideally, each test case is
    separate from the others; constructs such as mock objects can assist in separating unit tests. This type of testing is mostly done by the developers and not by end-users.
    Integration testing
    Integration testing can proceed in a number of different ways, which can be broadly characterized as top down or bottom up. In top down integration testing the high level control routines are tested first, possibly with the middle level control structures present only as stubs. Subprogram stubs were presented in Section 2 as incomplete subprograms which are only present to allow the higher level control routines to be tested. Thus a menu driven program may have the major menu options initially only present as stubs, which merely announce that they have been successfully called, in order to allow the high level menu driver to be tested.
    Top down testing can proceed in a depth-first or a breadth-first manner. For depth-first integration each module is tested in increasing detail, replacing more and more levels of detail with actual code rather than stubs. Alternatively breadth-first would proceed by refining all the modules at the same level of control
    throughout the application. In practice a combination of the two techniques would be used. At the initial stages all the modules might be only partly functional, possibly being implemented only to deal with non-erroneous data. These would be tested in breadth-first manner, but over a period of time each would be
    replaced with successive refinements which were closer to the full functionality. This allows depth-first testing of a module to be performed simultaneously with breadth-first testing of all the modules.The other major category of integration testing is bottom up integration testing where an individual module is
    tested from a test harness. Once a set of individual modules have been tested they are then combined into a collection of modules, known as builds, which are then tested by a second test harness. This process can continue until the build consists of the entire application.
    In practice a combination of top-down and bottom-up testing would be used. In a large software project being developed by a number of sub-teams, or a smaller project where different modules were being built by individuals. The sub-teams or individuals would conduct bottom-up testing of the modules which they were
    constructing before releasing them to an integration team which would assemble them together for top-down testing.
    I think this will help.
    Thanks ,
    Saptarshi

  • Unit Testing and Integrating testing In HR

    Dear Sap Gurus,
    Would you be kind enough to  give me an example of unit testing and integrating testing??  what do you test, eg..TC and what else.. what happened.??. And also an example of Integrating testing  ..and an example ....I know what unit and integrating test is ..and with a good example, i will have a great idea about it ...thanks a lot.

    Hi Pooja
    Unit Testing:
    A process for verifying that software, a system, or a system component performs its intended functions.
    Unit transactions are tested against their own specifications and design documents.
    Integration Testing
    An orderly progression of testing in which software elements, hardware elements or both are combined and tested until the entire system has been integrated.
    Integration tests deal mainly with the testing of cross-application process chains in addition to transactions and business processes. The process models and the test cases derived from these form the basis of these tests.
    Regards
    Vijay

  • Unit testing in nitrox

    I am new to eclipse and nitrox. Have been building webapps for years with IDEA.
    1) I find info in eclipse help concerning Junit based testing, I find nothing about unit testing in NitroX help. The webapps I am trying to import and build in NitroX must support unit testing. Eclipse Junit based help does not work for me in NitroX webapp.
    2) I have created several types of new webapps in NitroX accepting default parameters in wizard hoping to see how NitroX would lay out a new development environment. None of the results comes close to the development structure described in eclipse help. I find a src folder inside WEB-INF folder. This does not make sense to me and I can find no explanation of what NitroX expects with regards to project structure other than the existence of WEB-INF just below webapp context folder. My ultmate goal is to use both NitorX and IDEA accessing the same CVS repositories. So far I have imported several existing webapps via 'folder import' of source tree, but NitroX does note 'see' my testsrc or testbuild branches and I can not run tests except via my own custom Ant scripts. Any suggetions?
    Thanks

    1) I find info in eclipse help concerning Junit based testing, I find nothing about unit testing in NitroX help. The webapps I am trying to import and build in NitroX must support unit testing. Eclipse Junit based help does not work for me in NitroX webapp.
    If you have a Java class that extends JUnit, then in Eclipse, when you're in the Java-perspective, you can right-click on that class, "Run As..." and select "JUnit Test." That'll give you the JUnit Green or Red bar based on whether the tests succeeded/failed, etc.
    NitroX = Eclipse + Struts functionality. Anything you could do in Eclipse, you can do using NitroX.
    You could, for example, switch to the Java-perspective and perform JUnit testing exactly as in Eclipse without NitroX. Or, if you're in NitroX's "Web" perspective, you can go into Package Explorer and run JUnit tests from there.
    Re: 2)
    NitroX puts a "src" folder in your WEB-INF and sets that to be a source code folder, probably so when the application is deployed on the server, the source code is also on that server (in the isolated-from-the-web WEB-INF directory), which can be beneficial for several reasons, including your development environment and CVS server crashing.
    You could delete that folder, though, and use a folder you created yourself as a source code folder for that project. You'd just have to make sure that when you designate that source code folder (via: right-click on the project -> Properties -> Java Build Path -> Source Folders), you designate the output target of that folder to be /WEB-INF/classes. That'll make sure that, regardless of where your .java files are, your .class files will be placed in the appropriate folder inside your web application structure.
    Regarding NitroX not "see"ing your testsrc or testbuild branches (I assume you mean "folders" instead of "branches" ...), I'm assuming you're using NitroX's "Web" perspective, and looking for those folders in the AppXplorer window. Since your JUnit tests are not a part of the Struts application, you won't find them there -- instead, click over to the Package Explorer view, or switch over to Eclipse's Java Perspective, where everything will behave exactly as you'd expect.
    I hope this helps ..

Maybe you are looking for

  • A G4 Dying A Thousand Deaths

    I've been following a few threads here because I've had similar problems. I haven't seen a solution to this, however. I posted about frequent kernel panics, crashes and blue screens under BobMac, but when I went to log in for a further post the site

  • Razr Maxx no longer able to take screen shots

    Hi all, I've had a ton of trouble with the Razr Maxx. I've been through a few of the phones. First ICS update killed the phone, I've consistently had a Maxx that won't hold 4g connection in strong 4g areas and requires it to be put into airplane mode

  • Floating divs

    Hey TLF team, Do you have any immediate plans to implement floating blocks instead of just floating images?  I am currently implementing a stop-gap measure that splits the input up into multiple blocks, renders each to a different textflow, then lays

  • Protected Mode and incompatable system configurations

    After installing Adobe Reader 11.0.3 I get a message screen concerning "Protected Mode and incompatable system configurations". The screen gives me three options to continue, selecting any of these options still does not allow me to open any PDF. I'm

  • Problem with function module RS_VARIANT_CONTENTS

    Hi all, i´ve a problem by calling the FM RS_VARIANT_CONTENTS. Everytime I try to get the content of a variant the module crashes with the following dump: Call (PERFORM) to a non-existent routine. 025730     INSERT SCREEN_PROGS INDEX L_TABIX. 025740 0