Vous êtes sur la page 1sur 6

Example of Tool

Evaluation Criteria
Once the test engineer has narrowed the search for a particular type of test tool to
two or three lead candidates, the Evaluation Scorecard depicted in Table 1 can be
used to determine which of the tools best fits the particular requirements.

Tool Evaluation Scorecard

Table 1 Evaluation Scorecard—Automated GUI


Testing (Record/Playback) Tool
Weight Score Value
Test Tool Characteristic (1–10) (1–5) (1–50)
Ease of Use
Learning curve 7 5 35
Easy to maintain the tool 5 5 25
Easy to install—tool may not be used if difficult to install 5 3 15

Tool Customization
Can the tool be customized (can fields in tool be added 7 4 28
or deleted)?
Does the tool support the required test procedure 8 4 32
naming convention?

Platform Support
Can it be moved and run on several platforms at once, 8 4 32
across a network (that is, cross-Windows support, Win95,
and WinNT)?
continued

1
2 Example of Tool Evaluation Criteria

continued from page 1

Weight Score Value


Test Tool Characteristic (1–10) (1–5) (1–50)
Multiuser Access
What database does the tool use? Does it allow for scalability? 8 5 40
Network-based test repository—necessary when multiple 8 5 40
access to repository is required

Defect Tracking (For more detail


on evaluating defect tracking tools, see Chapter 8)
Does the tool come with an integrated defect-tracking feature? 10 3 30

Tool Functionality
Test scripting language—does the tool use a flexible, yet 9 5 45
robust scripting language? What is the complexity of the
scripting language: Is it 4 GL? Does it allow for modular
script development?
Complexity of scripting language 9 5 45
Scripting language allows for variable declaration and use; 9 5 45
allows passing of parameters between functions
Does the tool use a test script compiler or an interpreter? 9 5 45
Interactive test debugging—does the scripting language allow 8 4 32
the user to view variable values, step through the code,
integrate test procedures, or jump to other external
procedures?
Does the tool allow recording at the widget level (object 10 5 50
recognition level)?
Does the tool allow for interfacing with external .dll 9 5 45
and .exe files?
Published APIs—language interface capabilities 10 4 40
ODBC support—does the tool support any 10 4 40
ODBC-compliant database?
Is the tool intrusive (that is, does source code need to be 9 4 36
expanded by inserting additional statements)?
Communication protocols—can the tool be adapted to various 9 3 27
communication protocols (such as TCP/IP, IPX)?
Custom control support—does the tool allow you to map to 10 3 30
additional custom controls, so the tool is still compatible
and usable?
Ability to kick off scripts at a specified time; scripts can 9 5 45
run unattended
Allows for adding timers 10 5 50
Allows for adding comments during recording 7 5 35
continued
Example of Tool Evaluation Criteria 3

continued from page 2

Weight Score Value


Test Tool Characteristic (1–10) (1–5) (1–50)
Compatible with the GUI programming language and entire 10 5 50
hardware and software development environment used for
application under test (i.e., VB, Powerbuilder)
Can query or update test data during playback (that is, allows 10 4 40
the use of SQL statements)
Supports the creation of a library of reusable function 10 5 50
Allows for wrappers (shells) where multiple procedures can 10 5 50
be linked together and are called from one procedure
Test results analysis—does the tool allow you to easily see 10 3 30
whether the tests have passed or failed (that is, automatic
creation of test results log)?
Test execution on script playback—can the tool handle error 5 3 15
recovery and unexpected active windows, log the discrepancy,
and continue playback (automatic recovery from errors)?
Allows for synchronization between client and server 5 10 50
Allows for automatic test procedure generation 8 5 40
Allows for automatic data generation 8 5 40
Y2K compliance 10 5 50

Reporting Capability
Ability to provide graphical results (charts and graphs) 8 5 40
Ability to provide reports 8 5 40
What report writer does the tool use? 8 5 40
Can predefined reports be modified and/or can new 8 5 40
reports be created?

Performance and Stress Testing


Performance and stress testing tool is integrated 9 5 45
with GUI testing tool
Supports stress, load, and performance testing 10 3 30
Allows for simulation of users without requiring use of 10 3 30
physical workstations
Ability to support configuration testing (that is, tests can be 10 3 30
run on different hardware and software configurations)
Ability to submit a variable script from a data pool of library 10 3 30
of scripts/data entries and logon IDs/password
Supports resource monitoring (memory, disk space, 10 3 30
system resources)
Synchronization ability so that a script can access a record in 10 5 50
database at the same time to determine locking, deadlock
conditions, and concurrency control problems
continued
4 Example of Tool Evaluation Criteria

continued from page 3

Weight Score Value


Test Tool Characteristic (1–10) (1–5) (1–50)
Ability to detect when events have completed in a reliable 9 5 45
fashion
Ability to provide client to server response times 10 3 30
Ability to provide graphical results 8 5 40
Ability to provide performance measurements of data loading 10 5 50

Version Control
Does the tool come with integrated version control capability? 10 4 40
Can the tool be integrated with other version control tools 8 3 24

Test Planning and Management


Test planning and management tool is integrated with 8 5 40
GUI testing tool
Test planning and management tool is integrated with 8 5 40
requirements management tool
Test planning and management tool follows specific industry 7 4 28
standard on testing process (such as SEI/CMM, ISO)
Supports test execution management 10 5 50
Allows for test planning—does the tool support planning, 10 5 50
managing, and analyzing testing efforts? Can the tool
reference test plans, matrices, and product specifications
to create traceability?
Allows for measuring test progress 10 5 50
Allows for various reporting activities 9 4 36

Pricing
Is the price within the estimated price range? 10 4 40
What type of licensing is being used (floating, fixed)? 7 3 21
Is the price competitive? 9 4 36

Vendor Qualifications
Maturity of product 8 4 32
Market share of product 8 4 32
Vendor qualifications, such as financial stability and length 8 4 32
of existence. What is the vendor’s track record?
Are software patches provided, if deemed necessary? 8 4 32
Are upgrades provided on a regular basis? 8 5 40
Customer support 10 3 30
Training is available 9 4 36
Is a tool Help feature available? Is the tool well documented? 9 5 45
Availability and access to tool user groups 8 4 32
Total Value 2,638
Example of Tool Evaluation Criteria 5

As the weighted values for the test tool characteristics will vary with each type of test
tool, the test team may wish to develop an evaluation scorecard form for each type
of test tool required. In Table 1, an automated GUI test tool (capture/playback)
candidate is evaluated against the desired test tool characteristics. The total value of
2,638 for this candidate must then be compared with the total values derived for the
other two candidates. As noted in the sample scorecard summary below, Candidate
3 achieved a rating of 75.3% in being able to provide coverage for all the desired test
tool characteristics:

Candidate Score Rating

Candidate 1 2,360 67.4%


Candidate 2 2,530 72.3%
Candidate 3 2,638 75.3%

An optional evaluation scoring method involves sizing up the three candidates


using only the most important test tool characteristics. Note that 12 of the charac-
teristics were assigned a weight of 10. Table 2 reflects the scores for the three test
tool candidates using a preferred scorecard form based upon product information
obtained from each vendor.

Table 2 Preferred Scorecard—GUI Record/Playback Tool


Candidate 1 Candidate 2 Candidate 3
Test Tool Characteristic (1–5) (1–5) (1–5)
Integrated defect-tracking feature 3 5 3
Recording at the widget level 5 5 5
Published APIs—Language interface 4 5 4
capabilities
ODBC support—tool supports 4 4 4
ODBC-compliant databases
Custom control support 3 5 3
Allows for adding timers 4 5 5
Compatible with GUI language/ 5 5 5
development environment
Can query or update test data during 4 4 4
playback
Supports the creation of a library 4 4 5
of reusable functions
Allows for wrappers (shells) 4 5 5
Test results analysis 3 5 3
continued
6 Example of Tool Evaluation Criteria

continued from page 5

Candidate 1 Candidate 2 Candidate 3


Test Tool Characteristic (1–5) (1–5) (1–5)
Y2K compliance 5 5 5
Supports stress, load, and performance 3 5 3
testing
Allows for simulation of users 3 5 3
Supports configuration testing 3 3 3
Ability to use variable scripts 3 4 3
Supports resource monitoring 3 4 3
Synchronization ability 4 5 5
Client to server response times 3 4 3
Performance measurements of data loading 3 4 3
Version control 4 4 4
Supports test execution management 5 4 5
Allows for test planning 4 3 5
Measuring test progress 5 5 5
Price is within estimated range 3 5 4
Customer support 3 5 3
Total Value 970 11700 10300

A Preferred Scorecard Summary is provided in Table 2. Note that using this dif-
ferent model for scoring, test tool Candidate 2 achieves a higher rating than Candi-
date 3, which had posted the highest rating using the Evaluation Scorecard method.
Candidate 2 achieved a rating of 90.0% for being able to provide coverage for the
highest priority test tool characteristics.

Candidate Score Rating

Candidate 1 97 74.6%
Candidate 2 117 90.0%
Candidate 3 103 79.2%

Remember that the evaluation for each kind of test tool being considered for an
organization or project is different. Each particular type of test tool has its own par-
ticular desired characteristics and a different weight scheme for the tool’s desired
characteristics. The guidelines of what to look for and weigh when evaluating a
GUI test tool will be different from guidelines for how to evaluate a network mon-
itoring tool.

Vous aimerez peut-être aussi