You are required to read and agree to the below before accessing a full-text version of an article in the IDE article repository.
The full-text document you are about to access is subject to national and international copyright laws. In most cases (but not necessarily all) the consequence is that personal use is allowed given that the copyright owner is duly acknowledged and respected. All other use (typically) require an explicit permission (often in writing) by the copyright owner.
For the reports in this repository we specifically note that
- the use of articles under IEEE copyright is governed by the IEEE copyright policy (available at http://www.ieee.org/web/publications/rights/copyrightpolicy.html)
- the use of articles under ACM copyright is governed by the ACM copyright policy (available at http://www.acm.org/pubs/copyright_policy/)
- technical reports and other articles issued by M‰lardalen University is free for personal use. For other use, the explicit consent of the authors is required
- in other cases, please contact the copyright owner for detailed information
By accepting I agree to acknowledge and respect the rights of the copyright owner of the document I am about to access.
If you are in doubt, feel free to contact webmaster@ide.mdh.se
Dynamic Integration Test Selection Based on Test Case Dependencies
Publication Type:
Conference/Workshop Paper
Venue:
The 11th Workshop on Testing: Academia-Industry Collaboration, Practice and Research Techniques
Abstract
Prioritization, selection and minimization of test
cases are well-known problems in software testing. Test case
prioritization deals with the problem of ordering an existing set
of test cases, typically with respect to the estimated likelihood
of detecting faults. Test case selection addresses the problem of
selecting a subset of an existing set of test cases, typically by
discarding test cases that do not add any value in improving the
quality of the software under test. Most existing approaches for
test case prioritization and selection suffer from one or several
drawbacks. For example, they to a large extent utilize static
analysis of code for that purpose, making them unfit for higher
levels of testing such as integration testing. Moreover, they do not
exploit the possibility of dynamically changing the prioritization
or selection of test cases based on the execution results of prior
test cases. Such dynamic analysis allows for discarding test cases
that do not need to be executed and are thus redundant. This
paper proposes a generic method for prioritization and selection
of test cases in integration testing that addresses the above issues.
We also present the results of an industrial case study where
initial evidence suggests the potential usefulness of our approach
in testing a safety-critical train control management subsystem.
Bibtex
@inproceedings{Tahvili4298,
author = {Sahar Tahvili and Mehrdad Saadatmand and Stig Larsson and Wasif Afzal and Markus Bohlin and Daniel Sundmark},
title = {Dynamic Integration Test Selection Based on Test Case Dependencies},
month = {April},
year = {2016},
booktitle = {The 11th Workshop on Testing: Academia-Industry Collaboration, Practice and Research Techniques},
url = {http://www.es.mdu.se/publications/4298-}
}