Test
Cases for Testing Web and Desktop Applications – Comprehensive Testing
Checklist
|
This is
a testing checklist for web and desktop applications. Make testing checklist
as an integral part of test cases writing process. Using this checklist you
can easily create hundreds of test cases for testing web or desktop
applications. These are all general test cases and should be applicable for
almost all kind of applications. Refer these tests while writing test cases
for your project and I’m sure you will cover most testing types except the
application specific business rules provided in your SRS documents. Though
this is a common checklist, I recommend preparing a standard testing
checklist tailored to your specific needs using below test cases in addition
with application specific tests.
|
Importance
of Using Checklist for Testing:
|
-
Maintaining a standard repository of reusable test cases for your application
will ensure the most common bugs will be caught more quickly.
|
-
Checklist helps to quickly complete writing test cases for new versions of
the application.
|
-
Reusing test cases help to save money on resources to write repetitive tests.
|
-
Important test cases will be covered always making it almost impossible to
forget.
|
-
Testing checklist can be referred by developers to ensure most common issues
are fixed in development phase itself.
|
Few
notes to remember:
|
1)
Execute these scenarios with different user roles e.g. admin user, guest user
etc.
|
2) For
web applications these scenarios should be tested on multiple browsers like
IE, FF, Chrome, and Safari with versions approved by client.
|
3) Test
with different screen resolutions like 1024 x 768, 1280 x 1024, etc.
|
4)
Application should be tested on variety of displays like LCD, CRT, Notebooks,
Tablets, and Mobile phones.
|
4) Test
application on different platforms like Windows, Mac, Linux operating
systems.
|
Comprehensive
Testing Checklist for Testing Web and Desktop Applications:
|
Assumptions: Assuming that your application
supports following functionality
|
- Forms
with various fields
|
- Child
windows
|
-
Application interacts with database
|
-
Various search filter criteria and display results
|
- Image
upload
|
- Send
email functionality
|
- Data
export functionality
|
General
Test Scenarios
|
1. All
mandatory fields should be validated and indicated by asterisk (*) symbol
|
2.
Validation error messages should be displayed properly at correct position
|
3. All
error messages should be displayed in same CSS style (e.g. using red color)
|
4.
General confirmation messages should be displayed using CSS style other than
error messages style (e.g. using green color)
|
5. Tool
tips text should be meaningful
|
6.
Dropdown fields should have first entry as blank or text like ‘Select’
|
7.
Delete functionality for any record on page should ask for confirmation
|
8.
Select/deselect all records options should be provided if page supports
record add/delete/update functionality
|
9.
Amount values should be displayed with correct currency symbols
|
10.
Default page sorting should be provided
|
11.
Reset button functionality should set default values for all fields
|
12. All
numeric values should be formatted properly
|
13.
Input fields should be checked for max field value. Input values greater than
specified max limit should not be accepted or stored in database
|
14.
Check all input fields for special characters
|
15.
Field labels should be standard e.g. field accepting user’s first name should
be labeled properly as ‘First Name’
|
16.
Check page sorting functionality after add/edit/delete operations on any
record
|
17.
Check for timeout functionality. Timeout values should be configurable. Check
application behavior after operation timeout
|
18.
Check cookies used in an application
|
19.
Check if downloadable files are pointing to correct file paths
|
20. All
resource keys should be configurable in config files or database instead of
hard coding
|
21.
Standard conventions should be followed throughout for naming resource keys
|
22.
Validate markup for all web pages (validate HTML and CSS for syntax errors)
to make sure it is compliant with the standards
|
23.
Application crash or unavailable pages should be redirected to error page
|
24.
Check text on all pages for spelling and grammatical errors
|
25.
Check numeric input fields with character input values. Proper validation
message should appear
|
26.
Check for negative numbers if allowed for numeric fields
|
27.
Check amount fields with decimal number values
|
28.
Check functionality of buttons available on all pages
|
29. User
should not be able to submit page twice by pressing submit button in quick
succession.
|
30.
Divide by zero errors should be handled for any calculations
|
31.
Input data with first and last position blank should be handled correctly
|
GUI and
Usability Test Scenarios
|
1. All
fields on page (e.g. text box, radio options, dropdown lists) should be
aligned properly
|
2.
Numeric values should be right justified unless specified otherwise
|
3.
Enough space should be provided between field labels, columns, rows, error
messages etc.
|
4.
Scroll bar should be enabled only when necessary
|
5. Font
size, style and color for headline, description text, labels, infield data,
and grid info should be standard as specified in SRS
|
6.
Description text box should be multi-line
|
7.
Disabled fields should be grayed out and user should not be able to set focus
on these fields
|
8. Upon
click of any input text field, mouse arrow pointer should get changed to
cursor
|
9. User
should not be able to type in drop down select lists
|
10.
Information filled by users should remain intact when there is error message
on page submit. User should be able to submit the form again by correcting
the errors.
|
11.
Check if proper field labels are used in error messages
|
12.
Dropdown field values should be displayed in defined sort order
|
13. Tab
and Shift+Tab order should work properly
|
14.
Default radio options should be pre-selected on page load
|
15.
Field specific and page level help messages should be available
|
16.
Check if correct fields are highlighted in case of errors
|
17.
Check if dropdown list options are readable and not truncated due to field
size limit
|
18. All
buttons on page should be accessible by keyboard shortcuts and user should be
able to perform all operations using keyboard
|
19.
Check all pages for broken images
|
20.
Check all pages for broken links
|
21. All
pages should have title
|
22.
Confirmation messages should be displayed before performing any update or
delete operation
|
23. Hour
glass should be displayed when application is busy
|
24. Page
text should be left justified
|
25. User
should be able to select only one radio option and any combination for check
boxes.
|
Test
Scenarios for Filter Criteria
|
1. User
should be able to filter results using all parameters on the page
|
2.
Refine search functionality should load search page with all user selected
search parameters
|
3. When
there is at least one filter criteria is required to perform search
operation, make sure proper error message is displayed when user submits the
page without selecting any filter criteria.
|
4. When
at least one filter criteria selection is not compulsory user should be able
to submit page and default search criteria should get used to query results
|
5.
Proper validation messages should be displayed for invalid values for filter
criteria
|
Test
Scenarios for Result Grid
|
1. Page
loading symbol should be displayed when it’s taking more than default time to
load the result page
|
2. Check
if all search parameters are used to fetch data shown on result grid
|
3. Total
number of results should be displayed on result grid
|
4.
Search criteria used for searching should be displayed on result grid
|
5.
Result grid values should be sorted by default column.
|
6.
Sorted columns should be displayed with sorting icon
|
7.
Result grids should include all specified columns with correct values
|
8.
Ascending and descending sorting functionality should work for columns
supported with data sorting
|
9.
Result grids should be displayed with proper column and row spacing
|
10.
Pagination should be enabled when there are more results than the default
result count per page
|
11.
Check for Next, Previous, First and Last page pagination functionality
|
12.
Duplicate records should not be displayed in result grid
|
13.
Check if all columns are visible and horizontal scroll bar is enabled if
necessary
|
14.
Check data for dynamic columns (columns whose values are calculated
dynamically based on the other column values)
|
15. For
result grids showing reports check ‘Totals’ row and verify total for every
column
|
16. For
result grids showing reports check ‘Totals’ row data when pagination is
enabled and user navigates to next page
|
17.
Check if proper symbols are used for displaying column values e.g. % symbol
should be displayed for percentage calculation
|
18.
Check result grid data if date range is enabled
|
Test
Scenarios for a Window
|
1. Check
if default window size is correct
|
2. Check
if child window size is correct
|
3. Check
if there is any field on page with default focus (in general, the focus
should be set on first input field of the screen)
|
4. Check
if child windows are getting closed on closing parent/opener window
|
5. If
child window is opened, user should not be able to use or update any field on
background or parent window
|
6. Check
window minimize, maximize and close functionality
|
7. Check
if window is re-sizable
|
8. Check
scroll bar functionality for parent and child windows
|
9. Check
cancel button functionality for child window
|
Database
Testing Test Scenarios
|
1. Check
if correct data is getting saved in database upon successful page submit
|
2. Check
values for columns which are not accepting null values
|
3. Check
for data integrity. Data should be stored in single or multiple tables based
on design
|
4. Index
names should be given as per the standards e.g.
IND__
|
5.
Tables should have primary key column
|
6. Table
columns should have description information available (except for audit
columns like created date, created by etc.)
|
7. For
every database add/update operation log should be added
|
8.
Required table indexes should be created
|
9. Check
if data is committed to database only when the operation is successfully
completed
|
10. Data
should be rolled back in case of failed transactions
|
11.
Database name should be given as per the application type i.e. test, UAT,
sandbox, live (though this is not a standard it is helpful for database
maintenance)
|
12.
Database logical names should be given according to database name (again this
is not standard but helpful for DB maintenance)
|
13.
Stored procedures should not be named with prefix “sp_”
|
14.
Check is values for table audit columns (like createddate, createdby,
updatedate, updatedby, isdeleted, deleteddate, deletedby etc.) are populated
properly
|
15.
Check if input data is not truncated while saving. Field length shown to user
on page and in database schema should be same
|
16.
Check numeric fields with minimum, maximum, and float values
|
17.
Check numeric fields with negative values (for both acceptance and
non-acceptance)
|
18.
Check if radio button and dropdown list options are saved correctly in
database
|
19.
Check if database fields are designed with correct data type and data length
|
20.
Check if all table constraints like Primary key, Foreign key etc. are
implemented correctly
|
21. Test
stored procedures and triggers with sample input data
|
22.
Input field leading and trailing spaces should be truncated before committing
data to database
|
23. Null
values should not be allowed for Primary key column
|
Test
Scenarios for Image Upload Functionality (Also applicable for other file
upload functionality)
|
1. Check
for uploaded image path
|
2. Check
image upload and change functionality
|
3. Check
image upload functionality with image files of different extensions (e.g.
JPEG, PNG, BMP etc.)
|
4. Check
image upload functionality with images having space or any other allowed
special character in file name
|
5. Check
duplicate name image upload
|
6. Check
image upload with image size greater than the max allowed size. Proper error
message should be displayed.
|
7. Check
image upload functionality with file types other than images (e.g. txt, doc,
pdf, exe etc.). Proper error message should be displayed
|
8. Check
if images of specified height and width (if defined) are accepted otherwise
rejected
|
9. Image
upload progress bar should appear for large size images
|
10.
Check if cancel button functionality is working in between upload process
|
11.
Check if file selection dialog shows only supported files listed
|
12.
Check multiple images upload functionality
|
13.
Check image quality after upload. Image quality should not be changed after
upload
|
14.
Check if user is able to use/view the uploaded images
|
Test
Scenarios for Sending Emails
|
(Test
cases for composing or validating emails are not included)
|
(Make
sure to use dummy email addresses before executing email related tests)
|
1. Email
template should use standard CSS for all emails
|
2. Email
addresses should be validated before sending emails
|
3.
Special characters in email body template should be handled properly
|
4.
Language specific characters (e.g. Russian, Chinese or German language
characters) should be handled properly in email body template
|
5. Email
subject should not be blank
|
6.
Placeholder fields used in email template should be replaced with actual
values e.g. {Firstname} {Lastname} should be replaced with individuals first
and last name properly for all recipients
|
7. If
reports with dynamic values are included in email body, report data should be
calculated correctly
|
8. Email
sender name should not be blank
|
9.
Emails should be checked in different email clients like Outlook, Gmail,
Hotmail, Yahoo! mail etc.
|
10.
Check send email functionality using TO, CC and BCC fields
|
11.
Check plain text emails
|
12.
Check HTML format emails
|
13.
Check email header and footer for company logo, privacy policy and other
links
|
14.
Check emails with attachments
|
15.
Check send email functionality to single, multiple or distribution list
recipients
|
16.
Check if reply to email address is correct
|
17.
Check sending high volume of emails
|
Test
Scenarios for Excel Export Functionality
|
1. File
should get exported in proper file extension
|
2. File
name for the exported Excel file should be as per the standards
e.g. if file name is using timestamp, it should get replaced properly with
actual timestamp at the time of exporting the file
|
3. Check
for date format if exported Excel file contains date columns
|
4. Check
number formatting for numeric or currency values. Formatting should be same
as shown on page
|
5.
Exported file should have columns with proper column names
|
6.
Default page sorting should be carried in exported file as well
|
7. Excel
file data should be formatted properly with header and footer text, date,
page numbers etc. values for all pages
|
8. Check
if data displayed on page and exported Excel file is same
|
9. Check
export functionality when pagination is enabled
|
10.
Check if export button is showing proper icon according to exported file type
e.g. Excel file icon for xls files
|
11.
Check export functionality for files with very large size
|
12.
Check export functionality for pages containing special characters. Check if
these special characters are exported properly in Excel file
|
Performance
Testing Test Scenarios
|
1. Check
if page load time is within acceptable range
|
2. Check
page load on slow connections
|
3. Check
response time for any action under light, normal, moderate and heavy load
conditions
|
4. Check
performance of database stored procedures and triggers
|
5. Check
database query execution time
|
6. Check
for load testing of application
|
7. Check
for stress testing of application
|
8. Check
CPU and memory usage under peak load condition
|
Security
Testing Test Scenarios
|
1. Check
for SQL injection attacks
|
2.
Secure pages should use HTTPS protocol
|
3. Page
crash should not reveal application or server info. Error page should be
displayed for this
|
4.
Escape special characters in input
|
5. Error
messages should not reveal any sensitive information
|
6. All
credentials should be transferred over an encrypted channel
|
7. Test
password security and password policy enforcement
|
8. Check
application logout functionality
|
9. Check
for Brute Force Attacks
|
10.
Cookie information should be stored in encrypted format only
|
11.
Check session cookie duration and session termination after timeout or logout
|
11.
Session tokens should be transmitted over secured channel
|
13.
Password should not be stored in cookies
|
14. Test
for Denial of Service attacks
|
15. Test
for memory leakage
|
16. Test
unauthorized application access by manipulating variable values in browser
address bar
|
17. Test
file extension handing so that exe files are not uploaded and executed on
server
|
18.
Sensitive fields like passwords and credit card information should not have
auto complete enabled
|
19. File
upload functionality should use file type restrictions and also anti-virus
for scanning uploaded files
|
20.
Check if directory listing is prohibited
|
21.
Password and other sensitive fields should be masked while typing
|
22.
Check if forgot password functionality is secured with features like
temporary password expiry after specified hours and security question is
asked before changing or requesting new password
|
23.
Verify CAPTCHA functionality
|
24.
Check if important events are logged in log files
|
25.
Check if access privileges are implemented correctly
|
Below
are the standard fields of sample test case template:
|
Test
case ID: Unique
ID for each test case. Follow some convention to indicate types of test. E.g.
‘TC_UI_1′ indicating ‘user interface test case #1′.
|
Test
priority (Low/Medium/High): This is useful while test execution. Test
priority for business rules and functional test cases can be medium or higher
whereas minor user interface cases can be low priority. Test priority should
be set by reviewer.
|
Module
Name –
Mention name of main module or sub module.
|
Test
Designed By: Name of
tester
|
Test
Designed Date: Date
when wrote
|
Test
Executed By: Name of
tester who executed this test. To be filled after test execution.
|
Test
Execution Date: Date
when test executed.
|
Test
Title/Name: Test
case title. E.g. verify login page with valid username and password.
|
Test
Summary/Description:
Describe test objective in brief.
|
Pre-condition: Any prerequisite that must be
fulfilled before execution of this test case. List all pre-conditions in
order to successfully execute this test case.
|
Dependencies: Mention any dependencies on other
test cases or test requirement.
|
Test
Steps: List all
test execution steps in detail. Write test steps in the order in which these
should be executed. Make sure to provide as much details as you can. Tip – to
efficiently manage test case with lesser number of fields use this field to
describe test conditions, test data and user roles for running test.
|
Test
Data: Use of
test data as an input for this test case. You can provide different data sets
with exact values to be used as an input.
|
Expected
Result: What should be the system output after test
execution? Describe the expected result in detail including message/error
that should be displayed on screen.
|
Post-condition: What should be the state of the
system after executing this test case?
|
Actual
result: Actual
test result should be filled after test execution. Describe system behavior
after test execution.
|
Status
(Pass/Fail): If
actual result is not as per the expected result mark this test as failed.
Otherwise update as passed.
|
Notes/Comments/Questions: To support above fields if there are
some special conditions which can’t be described in any of the above fields
or there are questions related to expected or actual results mention those
here.
|
Add
following fields if necessary:
|
Defect
ID/Link: If test
status is fail, then include the link to defect log or mention the defect
number.
|
Test
Type/Keywords: This
field can be used to classify tests based on test types. E.g. functional,
usability, business rules etc.
|
Requirements: Requirements for which this test
case is being written. Preferably the exact section number of the requirement
doc.
|
Attachments/References:
This
field is useful for complex test scenarios. To explain test steps or expected
result using a visio diagram as a reference. Provide the link or location to
the actual path of the diagram or document.
|
Automation?
(Yes/No): Whether
this test case is automated or not. Useful to track automation status when
test cases are automated.
|
End of
Checklist
|