Dr.R.U.Anitha
1
Chapter 6 - Software Testing
Strategies
 Text Book: Software Engineering, Ian
Sommerville, 9TH
Edition , Pearson
2
Major Topics
 Verification and Validation
 Types of Testing
 System testing
3
Program testing
4
• Testing shows that a program does what it is intended
to do and to discover program defects before it is put
into use.
• It is a dynamic validation technique.
• To test, you execute a program using artificial data.
• Check the results of the test run for errors, anomalies or
information about the program’s non-functional
attributes.
• Can reveal the presence of errors NOT their absence.
• Testing is part of a more general verification and
validation process, which also includes static validation
Testing goals
 To demonstrate that software meets its requirements.
 For custom software, at least one test for every requirement in the
requirements document.
 For generic software products, have tests for all of the system features
AND feature combinations.
To discover situations in which the behavior of the software is incorrect
or undesirable.
 Defect testing is concerned with rooting out undesirable system
behavior such as system crashes, unwanted interactions with other
systems, incorrect computations and data corruption.
5
6
Validation and defect testing
• The first goal leads to validation testing
– It expect the system to perform correctly using a
given set of test cases that reflect the system’s
expected use.
• The second goal leads to defect testing
– The test cases are designed to expose defects. The test
cases in defect testing can be deliberately obscure and
need not reflect how the system is normally used.
7
Testing Process goals
• Validation testing
– To demonstrate to the developer and the system
customer that the software meets its requirements
– A successful test shows that the system operates as
intended.
• Defect testing
– To discover faults or defects in the software where its
behaviour is incorrect or not in conformance with its
specification
– A successful test is a test that makes the system perform
8
9
• Verification:
"Are we building the product right”.
• The software should conform to its specification.
• Validation:
"Are we building the right product”.
• The software should do what the user really requires.
Verification Vs validation
10
• Software inspections Concerned with analysis of the
static system representation to discover problems. (static
verification)
– May be supplement by tool-based document and code
analysis.
– Discussed in Chapter 15.
• Software testing Concerned with exercising and
observing product behaviour (dynamic verification)
– The system is executed with test data and its
operational behaviour is observed.
Inspections and testing
11
12
Software inspections
• These involve people examining the source representation
with the aim of discovering anomalies and defects.
• Inspections not require execution of a system so may be
used before implementation.
• They may be applied to any representation of the system
requirements, design,configuration data, test data, etc.).
• They have been shown to be an effective technique for
discovering program errors.
13
14
Stages of Testing
1. Development testing, where the system is tested
during development to discover bugs and defects.
2. Release testing, where a separate testing team test
a complete version of the system before it is
released to users.
3. User testing, where users or potential users of a
system test the system in their own environment.
1. Development Testing
• Development testing includes all testing activities that are
carried out by the team developing the system.
i. Unit testing, where individual program units or object
classes are tested. Unit testing should focus on testing the
functionality of objects or methods.
ii. Component testing, where several individual units are
integrated to create composite components. Component
testing should focus on testing component interfaces.
iii. System testing, where some or all of the components in
a system are integrated and the system is tested as a
whole. System testing should focus on testing component
interactions. 15
16
• Unit testing is the process of testing individual
components in isolation.
• It is a defect testing process.
• Units may be:
– Individual functions or methods within an object
– Object classes with several attributes and Methods
– Composite components with defined interfaces used
to access their functionality.
i. Unit testing
17
18
Testing strategies
• Partition testing, where it identify groups of inputs that
have common characteristics and should be processed in
the same way.
– It should choose tests from within each of these groups.
• Guideline-based testing, where it is use testing guidelines
to choose test cases.
– These guidelines reflect previous experience of the kinds
of errors that programmers often make when developing
components.
19
Unit Testing Guidelines
• Choose inputs that force the system to generate all
error messages
• Design inputs that cause input buffers to overflow
• Repeat the same input or series of inputs numerous
times
• Force invalid outputs to be generated
• Force computation results to be too large or too
small.
20
• Software components are often composite components that
are made up of several interacting objects.
– For example, in the weather station system, the
reconfiguration component includes objects that deal
with each aspect of the reconfiguration.
• It access the functionality of these objects through the
defined component interface.
• Testing composite components should therefore focus on
showing that the component interface behaves according to
its specification.
– It can assume that unit tests on the individual objects
within the component have been completed.
ii. Component Testing
21
Interface testing
• Objectives are to detect faults due to interface errors or
invalid assumptions about interfaces.
• Interface types
– Parameter interfaces Data passed from one method or
procedure to another.
– Shared memory interfaces Block of memory is shared
between procedures or functions.
– Procedural interfaces Sub-system encapsulates a set of
procedures to be called by other sub-systems.
– Message passing interfaces Sub-systems request services
from other sub-systems
22
Interface testing guidelines
• Design tests so that parameters to a called procedure
are at the extreme ends of their ranges.
• Always test pointer parameters with null pointers.
• Design tests which cause the component to fail.
• Use stress testing in message passing systems.
• In shared memory systems, vary the order in which
components are activated.
23
iii. System testing
• System testing during development involves integrating
components to create a version of the system and then
testing the integrated system.
• The focus in system testing is testing the interactions
between components.
• System testing checks that components are compatible,
interact correctly and transfer the right data at the right
time across their interfaces.
• System testing tests the emergent behaviour of a system.
Types of System Testing
A. Functional Testing
 Validates functional requirements
B. Performance Testing
 Validates non-functional requirements
C. Acceptance Testing
 Validates clients expectations
24
.
A. Functional Testing
Goal: Test functionality of system
 Test cases are designed from the requirements analysis
document (better: user manual) and centered around
requirements and key functions (use cases)
 The system is treated as black box
 Unit test cases can be reused, but new test cases have to
be developed as well.
25
B. Performance Testing
Goal: Try to violate non-functional requirements
 Test how the system behaves when overloaded.
 Can bottlenecks be identified? (First candidates for redesign in the next
iteration)
 Try unusual orders of execution
 Call a receive() before send()
 Check the system’s response to large volumes of data
 If the system is supposed to handle 1000 items, try it with 1001 items.
 What is the amount of time spent in different use cases?
 Are typical cases executed in a timely fashion?
26
Types of Performance Testing
 Stress Testing
 Stress limits of system
 Volume testing
 Test what happens if large
amounts of data are handled
 Configuration testing
 Test the various software and
hardware configurations
 Compatibility test
 Test backward compatibility
with existing systems
 Timing testing
 Evaluate response times and
time to perform a function
 Security testing
 Try to violate security
requirements
 Environmental test
 Test tolerances for heat,
humidity, motion
 Quality testing
 Test reliability, maintain-
ability & availability
 Recovery testing
 Test system’s response to
presence of errors or loss of
data
 Human factors testing
 Test with end users.
27
3. Acceptance Testing
 Acceptance testing, a testing technique performed to determine whether
or not the software system has met the requirement specifications.
 The main purpose of this test is to evaluate the system's compliance with
the business requirements and verify if it is has met the required
criteria for delivery to end users.
 There are various forms of acceptance testing:
 User acceptance Testing
 Business acceptance Testing
 Alpha Testing
 Beta Testing
28
2. Release testing
 Release testing is the process of testing a particular release of a
system that is intended for use outside of the development team.
 The primary goal of the release testing process is to convince the
customer of the system that it is good enough for use.
 Release testing, therefore, has to show that the system delivers its
specified functionality, performance and dependability, and that it
does not fail during normal use.
 Release testing is usually a black-box testing process where tests
are only derived from the system specification.
29
Release testing is a form of system testing. Important
differences:
 A separate team that has not been involved in the system
development, should be responsible for release testing.
 System testing by the development team should focus on
discovering bugs in the system (defect testing).
 The objective of release testing is to check that the system
meets its requirements and is good enough for external
use (validation testing).
30
 Requirements-based testing involves examining each
requirement and developing a test or tests for it.
 It is validation rather than defect testing: you are trying to
demonstrate that the system has properly implemented its
requirements.
 Scenario testing is an approach to release testing where you
devise typical scenarios of use and use these to develop test
cases for the system.
 Scenarios should be realistic and real system users should be
able to relate to them. If you have used scenarios as part of
the requirements engineering process, then you may be able
to reuse these as testing scenarios.
31
32
3. User testing
• User or customer testing is a stage in the testing process
in which users or customers provide input and advice on
system testing.
• User testing is essential, even when comprehensive
system and release testing have been carried out.
– The reason for this is that influences from the user’s
working environment have a major effect on the
reliability, performance, usability and robustness of a
system. These cannot be replicated in a testing
environment.
33
Types of user testing
• Alpha testing
– Users of the software work with the development
team to test the software at the developer’s site.
• Beta testing
– A release of the software is made available to users to
allow them to experiment and to raise problems that
they discover with the system developers.
• Acceptance testing
– Customers test a system to decide whether or not it is
ready to be accepted from the system developers and
deployed in the customer environment. Primarily for

Software Engineering CHAPTER 6 - Software Testing Strategies.ppt

  • 1.
  • 2.
    Chapter 6 -Software Testing Strategies  Text Book: Software Engineering, Ian Sommerville, 9TH Edition , Pearson 2
  • 3.
    Major Topics  Verificationand Validation  Types of Testing  System testing 3
  • 4.
    Program testing 4 • Testingshows that a program does what it is intended to do and to discover program defects before it is put into use. • It is a dynamic validation technique. • To test, you execute a program using artificial data. • Check the results of the test run for errors, anomalies or information about the program’s non-functional attributes. • Can reveal the presence of errors NOT their absence. • Testing is part of a more general verification and validation process, which also includes static validation
  • 5.
    Testing goals  Todemonstrate that software meets its requirements.  For custom software, at least one test for every requirement in the requirements document.  For generic software products, have tests for all of the system features AND feature combinations. To discover situations in which the behavior of the software is incorrect or undesirable.  Defect testing is concerned with rooting out undesirable system behavior such as system crashes, unwanted interactions with other systems, incorrect computations and data corruption. 5
  • 6.
    6 Validation and defecttesting • The first goal leads to validation testing – It expect the system to perform correctly using a given set of test cases that reflect the system’s expected use. • The second goal leads to defect testing – The test cases are designed to expose defects. The test cases in defect testing can be deliberately obscure and need not reflect how the system is normally used.
  • 7.
    7 Testing Process goals •Validation testing – To demonstrate to the developer and the system customer that the software meets its requirements – A successful test shows that the system operates as intended. • Defect testing – To discover faults or defects in the software where its behaviour is incorrect or not in conformance with its specification – A successful test is a test that makes the system perform
  • 8.
  • 9.
    9 • Verification: "Are webuilding the product right”. • The software should conform to its specification. • Validation: "Are we building the right product”. • The software should do what the user really requires. Verification Vs validation
  • 10.
    10 • Software inspectionsConcerned with analysis of the static system representation to discover problems. (static verification) – May be supplement by tool-based document and code analysis. – Discussed in Chapter 15. • Software testing Concerned with exercising and observing product behaviour (dynamic verification) – The system is executed with test data and its operational behaviour is observed. Inspections and testing
  • 11.
  • 12.
    12 Software inspections • Theseinvolve people examining the source representation with the aim of discovering anomalies and defects. • Inspections not require execution of a system so may be used before implementation. • They may be applied to any representation of the system requirements, design,configuration data, test data, etc.). • They have been shown to be an effective technique for discovering program errors.
  • 13.
  • 14.
    14 Stages of Testing 1.Development testing, where the system is tested during development to discover bugs and defects. 2. Release testing, where a separate testing team test a complete version of the system before it is released to users. 3. User testing, where users or potential users of a system test the system in their own environment.
  • 15.
    1. Development Testing •Development testing includes all testing activities that are carried out by the team developing the system. i. Unit testing, where individual program units or object classes are tested. Unit testing should focus on testing the functionality of objects or methods. ii. Component testing, where several individual units are integrated to create composite components. Component testing should focus on testing component interfaces. iii. System testing, where some or all of the components in a system are integrated and the system is tested as a whole. System testing should focus on testing component interactions. 15
  • 16.
    16 • Unit testingis the process of testing individual components in isolation. • It is a defect testing process. • Units may be: – Individual functions or methods within an object – Object classes with several attributes and Methods – Composite components with defined interfaces used to access their functionality. i. Unit testing
  • 17.
  • 18.
    18 Testing strategies • Partitiontesting, where it identify groups of inputs that have common characteristics and should be processed in the same way. – It should choose tests from within each of these groups. • Guideline-based testing, where it is use testing guidelines to choose test cases. – These guidelines reflect previous experience of the kinds of errors that programmers often make when developing components.
  • 19.
    19 Unit Testing Guidelines •Choose inputs that force the system to generate all error messages • Design inputs that cause input buffers to overflow • Repeat the same input or series of inputs numerous times • Force invalid outputs to be generated • Force computation results to be too large or too small.
  • 20.
    20 • Software componentsare often composite components that are made up of several interacting objects. – For example, in the weather station system, the reconfiguration component includes objects that deal with each aspect of the reconfiguration. • It access the functionality of these objects through the defined component interface. • Testing composite components should therefore focus on showing that the component interface behaves according to its specification. – It can assume that unit tests on the individual objects within the component have been completed. ii. Component Testing
  • 21.
    21 Interface testing • Objectivesare to detect faults due to interface errors or invalid assumptions about interfaces. • Interface types – Parameter interfaces Data passed from one method or procedure to another. – Shared memory interfaces Block of memory is shared between procedures or functions. – Procedural interfaces Sub-system encapsulates a set of procedures to be called by other sub-systems. – Message passing interfaces Sub-systems request services from other sub-systems
  • 22.
    22 Interface testing guidelines •Design tests so that parameters to a called procedure are at the extreme ends of their ranges. • Always test pointer parameters with null pointers. • Design tests which cause the component to fail. • Use stress testing in message passing systems. • In shared memory systems, vary the order in which components are activated.
  • 23.
    23 iii. System testing •System testing during development involves integrating components to create a version of the system and then testing the integrated system. • The focus in system testing is testing the interactions between components. • System testing checks that components are compatible, interact correctly and transfer the right data at the right time across their interfaces. • System testing tests the emergent behaviour of a system.
  • 24.
    Types of SystemTesting A. Functional Testing  Validates functional requirements B. Performance Testing  Validates non-functional requirements C. Acceptance Testing  Validates clients expectations 24
  • 25.
    . A. Functional Testing Goal:Test functionality of system  Test cases are designed from the requirements analysis document (better: user manual) and centered around requirements and key functions (use cases)  The system is treated as black box  Unit test cases can be reused, but new test cases have to be developed as well. 25
  • 26.
    B. Performance Testing Goal:Try to violate non-functional requirements  Test how the system behaves when overloaded.  Can bottlenecks be identified? (First candidates for redesign in the next iteration)  Try unusual orders of execution  Call a receive() before send()  Check the system’s response to large volumes of data  If the system is supposed to handle 1000 items, try it with 1001 items.  What is the amount of time spent in different use cases?  Are typical cases executed in a timely fashion? 26
  • 27.
    Types of PerformanceTesting  Stress Testing  Stress limits of system  Volume testing  Test what happens if large amounts of data are handled  Configuration testing  Test the various software and hardware configurations  Compatibility test  Test backward compatibility with existing systems  Timing testing  Evaluate response times and time to perform a function  Security testing  Try to violate security requirements  Environmental test  Test tolerances for heat, humidity, motion  Quality testing  Test reliability, maintain- ability & availability  Recovery testing  Test system’s response to presence of errors or loss of data  Human factors testing  Test with end users. 27
  • 28.
    3. Acceptance Testing Acceptance testing, a testing technique performed to determine whether or not the software system has met the requirement specifications.  The main purpose of this test is to evaluate the system's compliance with the business requirements and verify if it is has met the required criteria for delivery to end users.  There are various forms of acceptance testing:  User acceptance Testing  Business acceptance Testing  Alpha Testing  Beta Testing 28
  • 29.
    2. Release testing Release testing is the process of testing a particular release of a system that is intended for use outside of the development team.  The primary goal of the release testing process is to convince the customer of the system that it is good enough for use.  Release testing, therefore, has to show that the system delivers its specified functionality, performance and dependability, and that it does not fail during normal use.  Release testing is usually a black-box testing process where tests are only derived from the system specification. 29
  • 30.
    Release testing isa form of system testing. Important differences:  A separate team that has not been involved in the system development, should be responsible for release testing.  System testing by the development team should focus on discovering bugs in the system (defect testing).  The objective of release testing is to check that the system meets its requirements and is good enough for external use (validation testing). 30
  • 31.
     Requirements-based testinginvolves examining each requirement and developing a test or tests for it.  It is validation rather than defect testing: you are trying to demonstrate that the system has properly implemented its requirements.  Scenario testing is an approach to release testing where you devise typical scenarios of use and use these to develop test cases for the system.  Scenarios should be realistic and real system users should be able to relate to them. If you have used scenarios as part of the requirements engineering process, then you may be able to reuse these as testing scenarios. 31
  • 32.
    32 3. User testing •User or customer testing is a stage in the testing process in which users or customers provide input and advice on system testing. • User testing is essential, even when comprehensive system and release testing have been carried out. – The reason for this is that influences from the user’s working environment have a major effect on the reliability, performance, usability and robustness of a system. These cannot be replicated in a testing environment.
  • 33.
    33 Types of usertesting • Alpha testing – Users of the software work with the development team to test the software at the developer’s site. • Beta testing – A release of the software is made available to users to allow them to experiment and to raise problems that they discover with the system developers. • Acceptance testing – Customers test a system to decide whether or not it is ready to be accepted from the system developers and deployed in the customer environment. Primarily for

Editor's Notes

  • #24 When we are system testing, we are testing all subsystems together. The requirements have a large impact on the quality of system testing: The more explicit the requirements, the easier they are to test. We distinguish the following types of system testing Let’s walk through each of these system testing types
  • #25 Essentially the same as black box testing Unit test cases can be reused, but in end user oriented systems new test cases have to be developed as well.
  • #26 Push the (integrated) system to its limits.
  • #27 Stress Testing Stress limits of system (maximum # of users, peak demands, extended operation) Volume testing Test what happens if large amounts of data are handled Configuration testing Test the various software and hardware configurations Compatibility test Test backward compatibility with existing systems Security testing Try to violate security requirements