Model-based testing (MBT) in its most advanced form allows for automated test case identification, test data calculation, and test procedure generation from reference models describing the expected behaviour of the system under test (SUT). If the underlying algorithms for test case identification operate only on the syntactic representation of test models, however, the resulting test strength depends on the syntactic representation as well. This observation is true, even if syntactically differing models are behaviourally equivalent. In this paper, we present a systematic approach to elaborating test case selection strategies that only depend on the behavioural semantics of test models, but are invariant under syntactic transformations preserving the semantics. The benefits of these strategies are discussed, and practical generation algorithms are presented.
CITATION STYLE
Peleska, J., & Huang, W. L. (2016). Model-based testing strategies and their (In)dependence on syntactic model representations. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9933 LNCS, pp. 3–21). Springer Verlag. https://doi.org/10.1007/978-3-319-45943-1_1
Mendeley helps you to discover research relevant for your work.