Topic > Test Log: Test Plan and Test Case Management Software

Customer Requirements: This will be the list of requirements developed together with the customer and the developer. This list of requirements will constitute the "metric" against which the projects must be tested for acceptance. This part of the acceptance test log will be completed by the developer prior to the interview. Say no to plagiarism. Get a tailor-made essay on "Why Violent Video Games Shouldn't Be Banned"? Get an original essay Design evidence: This column should outline the evidence offered to demonstrate that technical and audience standards have been met within the designs. For example, wireframe documentation could be used as evidence that web pages fit the screen size of the user's device, and customer requirements documentation could be used as evidence that the developer intends to create a site using HTML 5 technologies and CSS. This part of the acceptance test log will be completed by the developer prior to the interview. Accepted: This is simply a yes or no column to quickly see if the particular requirement in question has been verified by the customer as an acceptable standard. This part of the acceptance test log will be completed by the developer during. Comments: This part of the table can be used to record customer feedback. It is hoped that the client agrees with your assessment of the projects and accepts that their standards have been met, but if not their views can be recorded here. That information can then be used later to assist in redesigns with the goal of revisiting customers and gaining their approval. This part of the acceptance test log will be completed by the developer during the interview. Date and signature: It should be at the end of the document where the client and developer can sign that the projects meet the expected technical and audience standards. This will protect the developer from a client changing their mind and will also protect the client from poor workmanship and partially completed projects. This part of the acceptance test log would be completed by the developer and the customer immediately after the interview. With the design documents in hand and the acceptance test log prepared with the Customer Requirements and Design Test columns completed, the developer would then seek a formal interview with the customer to validate the standards. This interview should record the client's response to each of the technical and audience standards outlined in the client requirements. If the client is unhappy with a particular part of the project and does not believe that standards would be met, this should be recorded so that developers can rework their designs before returning for further feedback. If the client is satisfied that all technical and public standards have been met, it would be prudent to have a signed copy of the acceptance record from the client and developer. This protects the developer from a customer changing their mind or disagreeing with the standards during the development phase; a prospect that will cost the developer time and money. Furthermore, it would also be prudent to have a legally binding contract signed by the developer and the client in which it is explicit that the project meets the technical and public standards to protect both parties in the project. Gray box testing is simply a fusion of black and white Box testing, where a developer can see the internal workings of the system, view the source code andalso view the outputs. Whichever testing method you choose to use, all tests must be recorded to demonstrate that the website has been verified and is fit for purpose. While you can purchase testing software, the simplest method is to create an Excel table with the appropriate headers and record each test as it is completed. A typical test log includes the following headers: Date: This is the date the test took place. This is important information since newer versions of the project may not have been tested, so knowing when testing took place can be helpful to a developer. Test Number: The test number should start at 1 and increment for each NEW test. However, if a particular test fails, a retest should be performed and the test number should indicate this. Typically, this is done by adding a letter or decimal number after the test number to show that it is a new test. For example, if test number 7 fails, the next test, a retest of number 7, should be labeled 7.1, or 7.01, or 7a, something appropriate to show the reader of the test log that it is a new test , not a different one. test fully. Purpose of the test – The purpose of the test should be explained briefly. For example, "Try the index.html hyperlink to the Contact Us page." This column should be short and sweet with just enough information to inform the reader about what is being tested. Test data: Some tests may require specific data input to test that the project can handle a variety of data. For example, if a developer created an HTML form to collect someone's name, they might want to run a couple of tests, one with valid test data like "kelvin" and another with invalid test data like "y4782oh42nlk-! ." Knowing specifically what test data was used in a specific test can often help a developer debug future issues. Some tests, however, do not have test data, such as testing whether a hyperlink works by simply clicking the hyperlink. In this case a developer can enter N/A (not applicable) or provide some form of detail such as "mouse click on hyperlink". Expected Results: Expected results are where the developer specifies EXACTLY what should happen if the test passes. Using my previous example of "Testing the index.html hyperlink to contact us page" as the test intent, a valid expected result would be "the contactus.html page loads in the same window". You should never write "works", this is not a valid expected result and appears rather lazy. Actual Results: This is where a developer will have actually tested and is now recording the test results. If a test passes, the actual results should be IDENTICAL to the expected ones. If the test fails, the developer should record what actually happened. So an actual valid result that passed the test would be “the contactus.html page loads in the same window”, which is identical to the expected result. A valid failed test might be "The contactus.html page failed to load, 404 error appears on screen." Comments: This is where a developer can comment on the test. The comments section is usually used when a developer has successfully re-ran a failed test and allows them to record what caused the test to fail and how they passed it. Different variations of this test plan can be found and a developer can add other columns if they see fit,.