<Your Project Name Here>
Test Plan
<Course, Semester, Year>
Modification history:
Version
|
Date
|
Who
|
Comment
|
v0.0
|
08/15/00
|
G. H. Walton
|
Template
|
v1.0
|
<date here>
|
<who>
|
<put comment to summarize the changes made in this
version>
|
...
|
|
|
|
Team Name: <your team name here>
Team Members:
- <person1 and that person's e-mail and web page -- name
and links>
- <person2 and that person's e-mail and web page -- name
and links>
- ...
Contents of this Document
- Introduction:
Overall Objective for Software
Test Activity
Reference Documents
- Description of Test Environment
- Overall Stopping Criteria
- Description of Individual
Test Cases
SECTION 1: Introduction
Reference Documents:
- Concept of Operations <include link here>
- Project Plan <include link here>
- SRS <include link here>
- <any other relevant documents; include full reference
information or link>
SECTION 2: Description of Test Environment
< What is the hardware and software in the environment in
which you will run the test? Who will be the testers (actual users? developers?
...) Will this test environment be the same environment in which the software
will operate? If not, how does it differ?>
SECTION 3: Stopping Criteria
<How will you determine when to stop testing the software
and either deliver it or send it "back" to development? Things to consider:
- If you find errors during testing: Will you stop testing
each time you find a problem and immediately fix that problem? Will you continue
testing and recording errors until you find a fatal error that won't allow
you to continue? Will you test for 2 hours and then fix whatever errors have
been found? Will you test for 2 hours and then hold a group meeting to decide
whether to continue to test? ...
- If you find no errors during testing: How many test cases
of what sort will you run before you declare the software to be "good enough
to deliver"?
- How do you define "good enough to deliver"? Does it require
that there are no known errors? Or no known errors other than cosmetic errors?
Or no known errors other than cosmetic errors and errors for which there is
a well-defined workaround? .... >
SECTION 4: Description of Individual
Test Cases
Describe EACH individual test to be run: (so if you plan to
run 20 test cases, you would answer the following questions for each of them.
Either a tabular format or a bulleted format is acceptable.)
- Test Objective: <exactly what are does this specific test
demonstrate?>
- Test Description: <exactly what will you test? What test
data will be used (specifically -- what data values, what data files? This
data must be determined in advance. So if you plan to use a test file, include
a link to the file here. If you plan to key in specific data, include the
data here. Make sure the data you choose will allow you to achieve your stated
objective for this test.>
- Test Conditions: <Under what conditions will you run this
test? This is relevant for software for which there are multiple "modes".
For some systems, the test conditions are totally described in the above test
environment section -- in this case, the response to this would be "See Test
Environment".
- Expected Results: <If the test executes correctly, what
will be the result -- i.e., exactly what will the output look like; what will
be the resulting data in the database, ...>
Template created by G. Walton (GWalton@mail.ucf.edu)
on March 28, 1999 and last modified on August 15, 2000.
This page last modified by <your name here>
(<your e-mail address text and link here> ) on <modification
date here>