Testing Software Product Lines Using Incremental Test Generation
|
|
|
- Leona Matthews
- 10 years ago
- Views:
Transcription
1 Testing Software Product Lines Using Incremental Test Generation Engin Uzuncaova Daniel Garcia Sarfraz Khurshid Don Batory Dept. of Electrical and Computer Engineering Dept. of Computer Sciences The University of Texas at Austin The University of Texas at Austin Abstract We present a novel specification-based approach for generating tests for products in a software product line. Given properties of features as first-order logic formulas, our approach uses SAT-based analysis to automatically generate test inputs for each product in a product line. To ensure soundness of generation, we introduce an automatic technique for mapping a formula that specifies a feature into a transformation that defines incremental refinement of test suites. Our experimental results using different data structure product lines show that incremental approach can provide an order of magnitude speed-up over conventional techniques. 1 Introduction The goal of software product lines is the systematic and efficient creation of products. Features are used to specify and distinguish products, where a feature is an increment in product functionality. Each product is defined by a unique combination of features. As product line technologies are applied to progressively more complex domains, the need for a systematic approach for product testing becomes more critical. Specification-based testing [6, 12] provides an effective approach for testing the correctness of software in general. The importance of using specifications in testing was realized at least three decades ago [12], and approaches based on specifications are widely used today. A typical approach generates test inputs using an input specification and checks the program using an oracle specification (correctness criteria). Several existing approaches can automatically generate test inputs from a specification as well as execute the program to check its outputs [15, 20]. For programs written in object-oriented languages, which are in common use for developing software product lines, a suitable specification language is Alloy [14] a declarative, first-order language based on relations. Alloy s relational basis and syntactic support for path expressions enable intuitive and succinct formulation of structurally complex properties of heap-allocated data structures, which pervade objectoriented programs. The Alloy Analyzer [13] a fully automatic tool based on propositional satisfiability solvers enables both test generation and correctness checking [15]. Given an Alloy formula that represents desired inputs, the analyzer solves the formula using a given bound on input size and enumerates the solutions. Test inputs are generated by translating each solution into a concrete object graph on which the program is executed. Correctness of the program is then checked using another Alloy formula that represents the expected relation between inputs and outputs. While the analyzer provides the necessary enabling technology for automated testing of programs with structurally complex inputs, test generation using the analyzer at present does not scale and is limited to generating small inputs (e.g., an object graph with less than ten nodes). To enable systematic testing of real applications we need novel approaches that scale to generation of larger inputs. The need is even greater for software product lines due to the current lack of support for analytical approaches for testing in this domain as well as due to the combinatorial nature of feature compositions [7]. This paper presents a novel approach for efficient test generation by combining ideas from software product lines and specification-based testing using Alloy. The novelty of our work is two-fold. First, each product is specified as a composition of features, where each feature is specified as an Alloy formula. An Alloy property of a program in a product line is thus specified as a composition (conjunction) of the Alloy formulas for each of the program s features. Second, the conventional use of the analyzer solves a complete specifica- 1
2 tion of a program to generate tests. Instead, we use the Alloy Analyzer to perform test generation incrementally; that is, we execute the analyzer more than once but on partial specifications, which are ideally easier problems to solve. To ensure soundness of generation, we introduce an automatic technique into our tool for mapping a formula that specifies a feature into a transformation that that defines incremental refinement of test suites. We present experimental results on a variety of data structure product lines that show incremental test generation can provide an order of magnitude speed-up over the conventional use. In a poster paper last year [24], we illustrated the use of Alloy for product line testing in the conventional fashion where tests are generated from a complete specification using a single execution of the analyzer. This paper makes the following new contributions: Incremental test generation. We introduce the notion of incremental generation of tests for testing products from a product line; Mapping. We define a mapping from a feature specification to a transformation among test suites and show how to perform it automatically; Implementation. Our prototype implementation uses the AHEAD and Alloy tool-sets to automate testing of product lines; and Evaluation. Experiments using a variety of data structure product lines that have intricate specifications show that our approach provides significant speed-ups over conventional techniques. 2 Example This section illustrates a simple product line of data structures. We use AHEAD [4] and Alloy [14] notations to explain our ideas. Section 5 presents a more sophisticated example. 2.1 A product line of binary trees Consider a family of binary trees. While all trees in this family are acyclic, they are differentiated on whether their nodes have parent pointers, or whether they have integer values satisfying constraints, or whether the trees cache the number of their nodes. The base product is an acyclic binary tree, which can be extended using a combination of three independent features: size, parent, and. We denote the collection of the base program and its features as an AHEAD model BT = {base, size, parent, }. A tree is defined by an expression. For example, the expression p = parent base, where denotes feature composition, defines a tree with parent pointers, p 0 size parent p 1 p 2 p 3 size size parent parent p 4 p 5 p 6 parent size p 7 p 0=base p 1=size base p 2=parent base Figure 1. Family of binary trees. p 3= base p 4=parent size base p 5= size base p 6= parent base p 7= parent size base and similarly, the expression s = base defines a binary tree(bst). Syntactically different expressions may be equivalent, e.g., size parent base = parent size base since size and parent are independent (i.e., commutative). Figure 1 characterizes the eight distinct products of the BT family. 2.2 Alloy annotated Jakarta code We next describe the basic class declarations and specifications that represent the BT family. The following annotated code declares the base classes: class BinaryTree { /*@ all n: root.*(left + right) n!in n.^(left + right) // no directed lone n.~(left + right) // at most one no n.left & n.right // left and right nodes Node root; } class Node { Node left, right; } A binary tree has a root node and each node has a left and a right child. The invariant annotation in comments states the class invariant, i.e., a constraint that a BinaryTree object must satisfy in any publicly visible state, such as a pre-state of a method execution [16]. The invariant is written as a universally quantified (keyword all) Alloy formula. The operator. represents relational composition; + is set union; and * is reflexive transitive closure. The expression root.*(left + right) represents the set of all nodes reachable from root following zero or more traversals along left or right edges. The invariant formula universally quantifies over all reachable nodes. It expresses three properties that are implicitly conjoined. (1) There are no directed cycles; (the operator! denotes negation and ^ denotes transitive closure; the keyword in represents set membership). (2) A node has at most one parent; (the operator ~ denotes relational transpose; the keyword lone represents a cardinality constraint of less than or equal to one on the corresponding set). (3) A node does not have another node as both its left child and its right child; (the operator & denotes set intersection). 2
3 AHEAD provides a veneer, Jakarta, on Java to facilitate development of product lines. The following Jakarta code uses the keyword refines, which denotes extension, to introduce the state that represents the feature size and the refinement of the invariant: refines class BinaryTree { /*@ refines size = #root.*(left + int size; } Note (1) the new field size in class Node and (2) the additional invariant that represents the correctness of size: the value of size field is the number of nodes reachable from root (inclusive). The Alloy operator # denotes cardinality of a set. When this refinement is applied to our original definition of BinaryTree, the size field is added to BinaryTree and the the new invariant is the conjunction of the original invariant with the size refinement. Similarly, we extend the base to introduce the state representing the feature parent by refining class BinaryTree and its invariant, and adding a new member to class Node: refines class BinaryTree { /*@ refines no all m, n: root.*(left + right) m in n.(left + right) <=> n = } refines class Node { Node parent; } The correctness of parent is: (1) root has no parent node (i.e., root.parent == null); and (2) if node m is the left or right child of node n then n is the parent of m and vice versa. We extend the base to introduce as follows. refines class BinaryTree { /*@ refines all n: root.*(left + right) all nl: n.left.*(left + right) { n.elem > nl.elem all nr: n.right.*(left + right) { n.elem < nr.elem } refines class Node { int element; } The constraint requires that the elements in the tree appear in the correct order: all elements in the left sub-tree of a node are smaller than its element and those in the right-subtree larger. s 0 s 3 τ τ τ t 0 Δ s Δ t Figure 2. BST commuting diagram. t Test generation We next illustrate how to generate inputs for methods defined in implementations of the products in the binary tree family. Since an input to a (public) method must satisfy its class invariant, we must generate valid inputs, i.e., inputs that satisfy the invariant. To illustrate, consider testing the size method in product p 5 = size base: // returns the number of nodes in the tree int size() {... } The method takes one input (the implicit input this). Generating a test input for method size requires solving p 5 s class invariant, i.e., acyclicity, size, and binary constraints (from Figure 1). Given the invariant in Alloy and a bound on the input size, the Alloy Analyzer can systematically enumerate all structures that satisfy the invariant; each structure represents a valid input for size (and other methods that take one tree as input). Given p 5 s invariant, the analyzer takes seconds on average to generate a tree with 10 nodes: This represents the conventional use of the analyzer. We use incremental solving to generate a desired test (Section 4). The commuting diagram in Figure 2 illustrates how our approach differs from the conventional approach. The nodes s i represent specifications for test generation for the corresponding products, e.g., s 0 represents the base specification the acyclicity constraint. The nodes t i represent the corresponding sets of test inputs. The horizontal arrow Δ s represents a refinement of the class invariant, i.e., the addition of constraints. The vertical arrows τ represent test generation using Alloy Analyzer. Δ t represents a transformation of tests for the base product into tests for base; Δ t is computed from Δ s and t 0 using the analyzer (Section 4). To generate tests t 3,the conventional approach follows the path τ Δ s. Our approach follows the alternative (but equivalent) path Δ t τ (dotted arrows). Given p 5 s invariant, we invoke the analyzer thrice. The total time it takes to generate a tree with exactly 10 nodes is 1.13 seconds on average, which is a 55 speed-up. Since our approach re-uses tests already generated for another product, when testing each product in a product line, the overall speed-up can be significantly greater. Detailed results are presented later in Section Feature Orientation A feature is an increment in program functionality. A software product-line (SPL) is a family of programs where no two programs have the same combination of 3
4 features 1. In the following sections, we sketch the basic ideas of feature orientation by a progression of models. 3.1 AHEAD GenVoca is a model of software product lines that is expressed purely in terms of transformations. Base programs are values (0-ary functions) and features are unary functions (i.e., transformations) that map programs to feature-extended programs, where denotes function composition: i x j x // adds feature i to program x // adds feature j to program x AHEAD extends GenVoca [3] by revealing the internal structure of GenVoca values and function as tuples. Every program has multiple representations: a program has source code, documentation, bytecode, makefiles, UML designs, etc. A GenVoca value is a tuple of representations. Base program f, for example, has a statechart model c f, a Java source code representation s f, and a Java bytecode representation b f. Program f s tuple is f = [c f, s f, b f ]. A GenVoca function maps a tuple of program representations to a tuple of updated representations. For example, feature j simultaneously refines f s statechart model (that adds new states and transitions, and extends existing actions), its source code (that adds new classes, new members to existing classes, and extends existing methods), and its bytecode (to execute j). If Δc j is the statechart refinement made by j, Δs j and Δb j are the corresponding refinements of source and bytecode, function j is the tuple j = [Δc j, Δs j, Δb j]. The representations of a program, such as p 1,are synthesized by tuple composition: p 1 = j f // GenVoca expression = [Δc j, Δs j, Δb j] [c f, s f, b f ] = [Δc j c f, Δs j s f, Δb j b f ] That is, the statechart of p 1 is produced by composing the base statechart with its refinement (Δc j c f ), the source code of p 1 s base with its refinement (Δs j s f ), etc. This is the essence of AHEAD [4]. 3.2 Feature-oriented model driven design AHEAD captures the lockstep refinement of program representations when a feature is composed with a program. However, AHEAD does not capture derivation relationships among program representations. Feature-Oriented Model Driven Development 1 Software product lines can be modeled in terms of functional and non-functional properties. Our approach focuses on functional properties as described in Section 3. Δc c j Δc h c i 2 c 3 φ φ φ Δs j Δs i s h javac b h Δb j s 2 javac s 3 javac Δb b i 2 b 3 Figure 3. Commuting Diagram. (FOMDD) generalizes AHEAD to include derivation relationships [23]. To illustrate, the relationship between Java source s f of program f and its bytecode b f is expressed by javac. Thatis, javac is a transformation that derives b f from s f. Similarly, one can imagine a transformation φ that derives the Java source s f from its statechart c f. Refinement and derivation relationships are expressed by a commuting diagram, where objects denote program representations, downward arrows represent derivations and horizontal arrows denote refinements. Figure 3 shows the commuting diagram for program p3 = i j h = [c 3, s 3, b 3]. 4 Our Approach This section describes our specification-based approach for test generation for systematic testing of implementations synthesized from an SPL. 4.1 FOMDD model For specification-based testing, the FOMDD models of our SPLs are defined as follows. Each program p of an SPL can be viewed as a pair: a specification s and a set of test inputs t, i.e., p = [s, t]. A feature f refines both a specification (Δs f ) and its test suite (Δt f ). In specification-based testing, the user provides s and Δs. To generate tests, we need a transformation τ that maps a specification s to its corresponding tests t. Also implementing test refinement Δt enables alternative techniques for test generation. We use the Alloy Analyzer to implement τ. In addition, we use the analyzer to implement transformation τ that automatically computes Δt: τ maps a test suite t and a specification refinement Δs to a corresponding test refinement Δt. Figure 2 shows the commuting diagram that corresponds to program p 0 = [s 0, t 0] composed with feature Objects An Alloy formula consists of a first-order logic constraint over primary variables (relations). An Alloy instance represents a valuation to these relations such that the formula evaluates to true. Mathematically, an 4
5 instance i is a function from a set of relations R to a power set of tuples 2 T where each tuple consists of indivisible atoms, i.e., i: R -> 2 T, where T is the set of all tuples in the bounded universe of discourse. Thus, for each Alloy relation, an instance gives a set of tuples that represents a value of the relation. Recall that to solve a formula, the Alloy Analyzer uses a scope that bounds the universe of discourse. The Kodkod back-end of the Alloy Analyzer [22] allows a scope to be specified using two bounds: a lower bound and an upper bound on the set of tuples that any valuation of a relation may take. Any instance must satisfy the following property: for every relation, each tuple in the lower bound must be present in the instance and no tuple that is not in the upper bound may be present in the instance. Mathematically, a bound b is a pair of two functions: a lower bound l and an upper bound u, each of type R -> 2 T. An instance can equivalently be viewed as bound b = [l, u] where l = u. Thus, in our model, a specification s is a pair of a formula f and a bound b, i.e., s = [f, b]; a test suite t is a set of instances. The specification refinement arrow Δs for specification s = [f, b] may refine the formula f or the bound b or both, i.e., and Δs = [Δf, Δb]. AHEAD s Jakarta notation provides the keyword refines to denote refinement. We overload this keyword to represent refinement of specifications. Refinement of a formula f transforms it into formula f Δf, where Δf represents the additional constraint. Refinement of a bound further restricts the lower or the upper bound or both. The transformation arrow τ represents test generation from the given specification. The test suite refinement arrow Δt enables an alternative test generation technique. The transformation arrow τ is a function from a test suite and a specification refinement to a test suite refinement. Implementing τ provides an implementation for Δt Paths In a commuting diagram, all paths that start at a desired specification and terminate at a desired test suite are equivalent, i.e., following any path gives the same test suite (up to isomorphism), in particular τ Δs = Δt τ. However, not all paths have the same associated cost, i.e., test generation along certain paths can be more efficient than others. Note that in the presence of feature interactions (Section 6), it may not be practical to traverse some Δt arrows. 4.2 Test generation Implementations of transformations τ and τ enable alternative techniques for test generation for products TestSuite τ (SpecificationRefinement Δs, TestSuite suite) { TestSuite suité = ; Formula formula = Δs.formula(); foreach (Test test: suite) { Bound bound = Δs.bound().update(test); suité = suité + Alloy.solve(formula, bound); } return suité ; } Figure 4. Test refinement algorithm. from a product line. The conventional use of the Alloy Analyzer allows a fully automatic implementation of τ: execute the analyzer on specification s and enumerate its instances. However, the conventional use of the analyzer restricts any path (in a commuting diagram) from a specification s to a test suite t to contain horizontal arrows that are labeled Δs only. This restriction requires performing transformation τ after all specification refinements have been performed, i.e., constraint solving is performed on the most complex of the specifications along any equivalent path. As specification formulas become more complex, execution of τ becomes more costly. For example, the analyzer takes one minute to generate an acyclic structure with 35 nodes. In contrast, the generation of an acyclic structure that also satisfies constraints with only 16 nodes does not terminate in 1 hour Algorithm We provide an algorithm (Figure 4), which enables a fully automatic implementation of the transformation τ. The algorithm assumes the monotonicity of feature semantics: when feature f is composed with base b, the resulting product s properties are a conjunction of b s properties and f s properties (Section 7). The impact of feature interactions on incremental test generation is discussed in Section 6. The algorithm takes as input a test suite t and a specification refinement Δs, and computes a new test suite, which refines the tests in t with respect to the constraints in Δs. The algorithm enables an incremental approach to test generation using successive applications of test refinement: to generate tests for a product that is composed of a base and a desired set of features, first generate a test suite for the base, and then iteratively refine the suite with respect to each of the features. In the specification-tests commuting diagram, we thus follow the path that starts with a vertical τ arrow and then consists solely of horizontal Δt arrows. Indeed, our algorithm also enables other paths 5
6 to be followed in the commuting diagram and hence it enables new approaches for test generation (Section 6). The algorithm transforms each test from the given suite into a test for the new suite. Incorporating the old test into the bound for the analyzer s guarantees the satisfaction of old constraints; in addition, the new solution includes valuations for the new relations introduced by the feature and satisfies the new constraints on these relations. Indeed, for features that constrain existing relations, the Alloy Analyzer may be unable to refine certain original tests, in which case the algorithm filters them out. In general, our algorithm τ implements an arbitrary relation from a given test suite (suite) and a specification refinement (Δs) to a desired test suite: (1) a particular test in suite may be refined into several new tests; and (2) certain tests in suite may not be refined and just ignored by the algorithm. A common case is when each test is refined to (at most) one test, i.e., τ is a (partial) function. Note that τ may not map two distinct tests onto the same new test (because the values of relations in original tests are not modified), i.e., τ is injective. Illustration. Consider the commuting diagram for binary trees (Figure 2). The following valuation represents a test input i from test suite t 0 for the base specification formula acyclic, as shown in Figure 5 (a): BinaryTree = { BT0 } Node = { N0, N1, N2 } root = { <BT0, N0> } left = { <N0, N1> } right = { <N0, N2>} Now consider transforming the test i into a test í for the specification formula of s 3, which represents acyclic. We run the analyzer on the formula and set the lower and upper bounds for BinaryTree, Node, root, left and right to the values in input i. The analyzer generates í by adding to the relations in i the new relations element and Int that models a set of integers: Int = { 0, 1, 2 } element = { <N0, 1> <N1, 0>, <N2, 2> } left root N 0 right N 1 N 2 (a) left root N 0 1 right N 1 0 N 2 2 (b) Figure 5. Test inputs. (a) An acyclic binary tree. (b) An acyclic binary tree with elements 0, 1, and 2. Figure 5 (b) graphically illustrates this tree, which is indeed a binary tree. Correctness. We next argue the soundness and completeness (with respect to the given input bounds) of our approach. We outline a simple induction argument. Consider generating tests for product p n = f n... f 1 f 0, where f 0 is a base product and each f i (i>0) is a feature. The induction base case holds trivially since the tests for the base are generated using a direct application of the Alloy Analyzer. For the induction step, consider generating test suite t k+1 for product p k+1 using test suite t k for product p k, where t k consists of exactly all the valid tests for p k. The soundness follows from the fact that the invocation of the analyzer does not change any values of relations that appear in p k. Thus, constraints for p k continue to be satisfied. Moreover, since the analyzer directly solves the constraints in the specification refinement, any solution it generates satisfies the additional constraints of p k+1 by definition. Thus, if the invocation of the analyzer returns a solution, it satisfies all constraints for p k+1. (Indeed, some tests for p k may simply be filtered out.) The completeness follows from the monotonicity of feature semantics: any valid test input for a product must satisfy properties of all its features. Let i k+1 be an arbitrary valid test input for p k+1. Let i k be an input that has the same values as i k+1 for all relations in p k and contains no other values for any relation. Then by the monotonicity property, i k is a valid input for p k. Thus, by the induction hypothesis, i k t k. Therefore, the foreach loop performs an iteration that refines i k. Since the analyzer enumerates all solutions, i k can spawn several new inputs and the output of the solver includes all of them. Thus, one of the solutions returned by its invocation must be i k+1 (up to isomorphism). Hence, i k+1 is generated by the algorithm. Therefore, all valid inputs for p k+1 are generated. 5 Evaluation The section presents an evaluation of our incremental approach to test generation using two subject product lines: binary trees and intentional names [1]. Section 2 introduced the binary tree product line. Section 5.1 describes the intentional naming product line. We tabulate and discuss the results for enumerating test inputs using the conventional approach and our incremental approach (Section 5.2). All experiments were performed on a 1.8GHz Pentium M processor using 512MB of RAM. All SAT formulas were solved using MiniSat [11]. Our tool Kesit uses the Java API of the Kodkod back-end [22] of the Alloy Analyzer. 6
7 conventional incremental product total refine- time speed vars clause time ment vars clause ref total up Binary Search Tree (scope=10) base n/a n/a n/a n/a n/a n/a size base size parent base parent base parent size base size parent size base size parent base parent parent size base size parent INS (scope=16) base n/a n/a n/a n/a n/a n/a attr-val base attr val label attr-val base attr val label record label attr-val base attr-val label record Table 1. Performance results for the subject product lines. All times are in milliseconds. 5.1 Intentional naming The Intentional Naming System (INS) [1] is a resource discovery architecture for dynamic networks. INS is implemented in Java; the core naming architecture is about 2000 lines of code. In previous work [15], we modeled INS in Alloy and discovered significant bugs in its design and implementation. Here, we show how incremental test generation gives a significant speed-up over the conventional approach. INS allows describing services using their properties. This enables client applications to state what service they want without having to specify where in the network topology it resides. Service properties in INS are described using intentional names, which are implemented using name-specifiers hierarchical arrangements of alternating levels of attributes and values. Attributes classify objects. Each attribute has a value that further classifies the object. A wildcard may be used if any value is acceptable. An attribute together with its value form an av-pair; each av-pair has a set of child av-pairs. The av-pairs form a tree structure. Services advertise themselves to name resolvers that maintain a database to store mappings between name-specifiers and name records, which include information about the current service locations. To test the correctness of key INS algorithms, we must generate advertisements and queries as test inputs. We differentiate each product in the intentional name product line based on whether there are attribute and value nodes, or whether attributes and values have labels satisfying the constraints for a name-specifier, or whether the trees have pointers from their leaf valuenodes to name-records. The following AHEAD model describes this family: INS = {base, attr-val, label, record}. Actual Alloy model for INS is not included due to space considerations. 5.2 Results Table 1 presents the experimental results for the two subject product lines. Conventional approach is test generation with the latest Alloy tool-set, whereas incremental refers to our approach Kesit. For each product, we tabulate the number of primary variables, the number of CNF clauses and the total time for the conventional approach. We also tabulate the number of additional Boolean variables, the number of additional CNF clauses, the additional time taken to refine previously generated tests and the total time for our incremental approach. The last column shows the speed-up. We generated 100 test inputs for each product and the tabulated times represent the average time to generate a single test for the product. We tabulate results for binary trees for 10 nodes and for intentional names for 16 nodes; these scopes are representative of the general characteristics we have observed during the experiments. As mentioned earlier, a product can be generated following different paths in the corresponding commuting diagrams: For each product, we show the results for which Kesit most significantly outperformed the traditional approach. Experiments show that Kesit can provide a speedup of over 66. However, it does not always provide a speed-up and for some products, such as size base and parent base, we observe a slow down in compari- 7
8 son with the conventional approach. While we expect SAT problems with fewer primary variables to be easier to solve, we observe that applying our algorithm to refinements that involve simple constraints introduces an excessive overhead. Therefore, the conventional approach seems to be more efficient for simple refinements. However, for more complex constraints, such as, our incremental approach performs significantly better. Parallel to that, as the scope increases, the performance improvement Kesit provides becomes more significant not only for complex constraints but also for the simple ones. The experiment results pertaining larger scopes are not presented in this paper due to space considerations. We obtain the highest speed-up for the refinement in the Binary Tree subject. With the conventional approach, going beyond the scope of 12 seems infeasible. Our incremental approach enables SAT solvers to handle significantly larger scopes because the resulting SAT problems are much simpler. For example, generating test cases for binary tree with the constraints involves clauses in the conventional approach, but Kesit works with only and 4773 clauses for the base and features respectively. We observe this effect with the INS model too. The number of primary variables and clauses are greater (i.e., 1128 and respectively) for the conventional approach due to the complexity and size of the complete model. However, incremental generation reduces the problem to two smaller refinements, attr-val and label, which involve smaller numbers of variables and clauses. To summarize, a key strength of Kesit is to solve more complex problems and reach larger scopes. There are two key findings that we have observed during our experiments: (1) for simple refinements, Kesit s performance is comparable to the conventional approach, and (2) for complex refinements, Kesit significantly outperforms the conventional approach. path time #filtered conv ρ 1: τ balance base ρ 2: τ balance base incr ρ 3: balance τ base (basic) ρ 4: balance τ base incr ρ 5: τ balance base (mixed) ρ 6: balance τ base results are averaged over 50 inputs Table 2. Comparison of different paths. The path ρ 5 is optimal. 6 Future Work 6.1 Further Optimizations FOMDD suggests that the conventional and incremental approaches are only two of many other approaches for generating tests, and that a combination of conventional and incremental may in fact be more efficient. Figure 6 illustrates a three-dimensional commuting diagram for balanced binary trees as described earlier. Our incremental approach is represented by a pair of paths in this cube, starting from base specification s 0 to test t 3 that first descends and then walks the bottom of the cube: balance τ base balance τ base The conventional approach, in contrast, follows a different set of paths that walks the top of the cube before descending: τ balance base τ balance base Clearly, there are other paths, and among them is a more efficient generation strategy. The path with bold arrows reflects this alternative strategy, where we first solve for base and balance constraints together and then incrementally solve for constraints: τ balance base The reason why this alternative path is more efficient is that many solutions of the base program are discarded when additional constraints (e.g., or balance) are added. It is actually cheaper to start with a slightly more complex specification, generate and extend its solutions, than starting from the base. We discovered this optimal path by examining all paths (see Table 2) [23]. Note that the number of paths that are filtered (meaning that the number of solutions that are subsequently discarded as they do not extend to solutions of more complex programs) is an important indicator of a path s performance. We s 1 balance s 0 s 2 t 1 t balance 0 t 2 balance balance Figure 6. Specification-tests commuting for BST. The path with bold arrows is ρ 5: τ balance base. s 3 t 3 8
9 recently developed a constraint prioritization approach that can assist in identifying an optimal path for test generation; details are described elsewhere [25]. 6.2 Feature Interactions While features often represent additional program functionality as we assumed in incremental test generation, one of the key issues in feature-based development is accounting for feature interactions [17], where features may replace existing functionality. Although it has been the subject of a large body of re [5], much about feature interactions is still not well understood. FOMDD allows features to have a more profound impact on properties than can be expressed by conjunction. An interaction occurs when a feature replaces (not just extends) existing constraints. In general, features can transform a property of a program (such as replacing an existing constraint with another, thus disrupting the monotonic increase that we assumed earlier), in which case our incremental approach may not apply directly. Certainly, an incremental approach could apply to feature compositions from the point of the last non-monotonic (i.e., last property-replacing) feature. And it might apply if replaced properties are simply removed from earlier features in order to emulate monotonic compositions. In any case, this will be an interesting subject for future re. 7 Related Work Nebut et al. [7] states that software product line processes still lack support for testing end-products using methods and techniques that are based on specific features of a product line, i.e., commonality and variability. While classical testing approaches can be applied in the product line domain, the very nature of feature composition and the large number of possible product configurations introduce a serious challenge for scalability. Much of the literature in testing software product lines focuses on planning and assessment of software testing [10, 18]. Our approach introduces an opportunity for tailoring the practices from the classical testing domain with respect to the specific requirements of software product lines. Approaches for regression testing [19] bear similarities to our work. A key problem addressed by these approaches is of test selection: a code-based selection technique attempts to identify a subset of existing tests that are likely to reveal faults in the modified program. Our incremental generation contrasts with test selection since we do not select tests from a given suite but instead we refine given tests using constraint solving. Barrett et al. [2] present an incremental approach for translating first-order logic formulas into SAT problems. Instead of translating the entire formula up front, they translate it incrementally as the is conducted by the SAT solver. This approach deals with the SAT solver semantics and interacts directly with the solver. In contrast, our approach works at a higher level and manipulates Alloy formulas. The two approaches are thus complementary and can be used in conjunction to further optimize test generation. Recently Cohen et al. [9] investigated the use of incremental satisfiability solvers for generating interaction test suites. Their algorithm uses the incremental solver MiniSAT [11] to optimize the AETG [8] test generation algorithm. Incremental SAT solvers have a direct application for incremental test generation and we plan to explore their use in testing product lines. In previous work, we developed the TestEra [15] framework for specification-based testing of Java programs. TestEra generates inputs and checks program correctness using Alloy specifications. TestEra performs bounded exhaustive testing. TestEra s generation enabled achieving high code coverage for unit testing of library code. TestEra discovered subtle bugs in stand-alone applications including a fault-tree analyzer [21]. This paper shows how we build on TestEra to provide a significantly more efficient test generator, Kesit, for testing software product lines using an incremental approach. Kesit not only generates tests more efficiently than TestEra, but also scales to generation of larger inputs, which enables novel strategies for software testing. To illustrate, bounded exhaustive testing can be complemented by testing on selected larger inputs that are generated using the same constraints. While this paper focuses on the use of Alloy for test generation, Alloy has various other applications, including modeling and checking of designs of software artifacts [14], and static analysis of code [26]. We believe our approach will enable new efficient analyses to support checking of not only the implementations but also the designs of product lines. 8 Conclusions Testing software product lines is an important and difficult problem. We presented a novel technique that incrementally generates tests for product lines, in general, and FOMDD models of product lines in particular. Our key insight to test generation comes from the definition of a feature: an increment in program functionality. We introduced an automatic technique for mapping a formula that specifies a feature into a transformation that defines incremental refinement of 9
10 test suites. Our approach performs test generation incrementally. The experimental results with our prototype Kesit shows that incremental test generation provides significant performance improvements over conventional means of test input generation. We believe incremental approaches hold much promise, not just in the context of product lines but also in the more general software testing context, e.g., for refining tests for regression testing. Acknowledgments This work was funded in part by NSF s Science of Design Awards #CCF , #CCF and #IIS References [1] W. Adjie-Winoto, E. Schwartz, H. Balakrishnan, and J. Lilley. The design and implementation of an intentional naming system. In Proc. of the 17th Int l Conference on Verification, Model Checking and Abstract Interpretation, Kiawah Island, SC, December [2] C. W. Barrett, D. L. Dill, and A. Stump. Checking satisfiability of first-order formulas by incremental translation to sat. In Proc. of the 14th Int l Conference on Computer Aided Verification, July [3] D. Batory and S. O Malley. The design and implementation of hierarchical software systems with reusable components. Comm. of the ACM, 1(4): , [4] D. Batory, J. N. Sarvela, and A. Rauschmayer. Scaling step-wise refinement. IEEE Transactions on Software Engineering, 30(6): , June [5] M. Calder, M. Kolberg, E. H. Magill, and S. Reiff- Marganiec. Feature interaction: a critical review and considered forecast. Computer Networks, 41(1): , [6] J. Chang and D. J. Richardson. Structural specification-based testing: Automated support and experimental evaluation. In Proc. of the 7th ACM SIGSOFT Symposium on the Foundations of Software Engineering, September [7] Y. L. T. Clémentine Nebut and J.-M. Jézéquel. System testing of product lines: From requirements to test cases. In Software Product Lines - Re Issues in Engineering and Management, pages Springer, [8] D.M.Cohen,S.R.Dalal,M.L.Fredman,andG.C. Patton. The AETG system: An approach to testing based on combinatorial design. IEEE Transactions on Software Engineering, 23(7): , [9] M. B. Cohen, M. B. Dwyer, and J. Shi. Exploiting constraint solving history to construct interaction test suites. Testing: Academic and Industrial Conference Practice and Re Techniques - MUTATION, pages , Sept [10] M. B. Cohen, M. B. Dwyer, and J. Shi. Coverage and adequacy in software product line testing. In Proc. of the ISSTA 2006 workshop on Role of software architecture for testing and analysis, NY, July [11] N. Een and N. Sorensson. An extensible sat-solver. In Proc. of the 6th Int l Conference on Theory and Applications of Satisfiability Testing, Italy, May [12] J. Goodenough and S. Gerhart. Toward a theory of test data selection. IEEE Transactions on Software Engineering, June [13] D. Jackson. Alloy: A lightweight object modeling notation. ACM Transactions on Software Engineering and Methodology, 11(2), April [14] D. Jackson. Software Abstractions: Logic, Language and Analysis. The MIT Press, Cambridge, MA, [15] S. Khurshid. Generating Structurally Complex Tests from Declarative Constraints. PhD thesis, Dept. of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, December [16] B. Liskov and J. Guttag. Program Development in Java: Abstraction, Specification, and Object-Oriented Design. Addison-Wesley, [17] J. Liu, D. Batory, and S. Nedunuri. Modeling interactions in feature oriented software designs. In Proc. of the the 8th Int l Conference on Feature Interactions in Telecommunications and Software Systems, [18] H. Muccini and A. van der Hoek. Towards testing product line architectures. In Proc. of the Int l Workshop on Test and Analysis of Component-Based Systems, Warsaw, Poland, April [19] G. Rothermel and M. J. Harrold. Analyzing regression test selection techniques. IEEE Transactions on Software Engineering, 22(8): , [20] K. Sen, D. Marinov, and G. Agha. CUTE: A concolic unit testing engine for C. In Proc. of the 13th ACM SIGSOFT Symposium on the Foundations of Software Engineering, Lisbon, Portugal, September [21] K. Sullivan, J. Yang, D. Coppit, S. Khurshid, and D. Jackson. Software assurance by bounded exhaustive testing. In Proc. of the Int l Symposium on Software Testing and Analysis, [22] E. Torlak and D. Jackson. Kodkod: A relational model finder. In Proc. of the 13th Int l Conference on Tools and Algorithms for Construction and Analysis of Systems, Braga, Portugal, March [23] S. Trujillo, D. Batory, and O. Diaz. Feature oriented model driven development: A case study for portlets. In Proc. of the 29th Int l Conference on Software Engineering, Minneapolis, MN, May [24] E. Uzuncaova, D. Garcia, S. Khurshid, and D. Batory. A specification-based approach to testing software product lines. In Proc. of the 15th ACM SIG- SOFT Symposium on the Foundations of Software Engineering, Croatia, September Poster Paper. [25] E. Uzuncaova and S. Khurshid. Constraint prioritization for efficient analysis of declarative models. In Proc. of the 15th Int l Symposium on Formal Methods, Turku, Finland, May [26] M. Vaziri. Finding Bugs Using a Constraint Solver. PhD thesis, Computer Science and Artificial Intelligence Lab, MIT,
Random vs. Structure-Based Testing of Answer-Set Programs: An Experimental Comparison
Random vs. Structure-Based Testing of Answer-Set Programs: An Experimental Comparison Tomi Janhunen 1, Ilkka Niemelä 1, Johannes Oetsch 2, Jörg Pührer 2, and Hans Tompits 2 1 Aalto University, Department
Optimizing Generation of Object Graphs in Java PathFinder
Optimizing Generation of Object Graphs in Java PathFinder Milos Gligoric, Tihomir Gvero, Steven Lauterburg, Darko Marinov, Sarfraz Khurshid JPF Workshop 1.5.8 Bugs Six Decades Ago 1947: Harvard Mark II
Verifying Semantic of System Composition for an Aspect-Oriented Approach
2012 International Conference on System Engineering and Modeling (ICSEM 2012) IPCSIT vol. 34 (2012) (2012) IACSIT Press, Singapore Verifying Semantic of System Composition for an Aspect-Oriented Approach
Component visualization methods for large legacy software in C/C++
Annales Mathematicae et Informaticae 44 (2015) pp. 23 33 http://ami.ektf.hu Component visualization methods for large legacy software in C/C++ Máté Cserép a, Dániel Krupp b a Eötvös Loránd University [email protected]
Formal Engineering for Industrial Software Development
Shaoying Liu Formal Engineering for Industrial Software Development Using the SOFL Method With 90 Figures and 30 Tables Springer Contents Introduction 1 1.1 Software Life Cycle... 2 1.2 The Problem 4 1.3
Software Modeling and Verification
Software Modeling and Verification Alessandro Aldini DiSBeF - Sezione STI University of Urbino Carlo Bo Italy 3-4 February 2015 Algorithmic verification Correctness problem Is the software/hardware system
µz An Efficient Engine for Fixed points with Constraints
µz An Efficient Engine for Fixed points with Constraints Kryštof Hoder, Nikolaj Bjørner, and Leonardo de Moura Manchester University and Microsoft Research Abstract. The µz tool is a scalable, efficient
Keywords: Regression testing, database applications, and impact analysis. Abstract. 1 Introduction
Regression Testing of Database Applications Bassel Daou, Ramzi A. Haraty, Nash at Mansour Lebanese American University P.O. Box 13-5053 Beirut, Lebanon Email: rharaty, [email protected] Keywords: Regression
Towards a Framework for Differential Unit Testing of Object-Oriented Programs
Towards a Framework for Differential Unit Testing of Object-Oriented Programs Tao Xie 1 Kunal Taneja 1 Shreyas Kale 1 Darko Marinov 2 1 Department of Computer Science, North Carolina State University,
npsolver A SAT Based Solver for Optimization Problems
npsolver A SAT Based Solver for Optimization Problems Norbert Manthey and Peter Steinke Knowledge Representation and Reasoning Group Technische Universität Dresden, 01062 Dresden, Germany [email protected]
Generating models of a matched formula with a polynomial delay
Generating models of a matched formula with a polynomial delay Petr Savicky Institute of Computer Science, Academy of Sciences of Czech Republic, Pod Vodárenskou Věží 2, 182 07 Praha 8, Czech Republic
From Workflow Design Patterns to Logical Specifications
AUTOMATYKA/ AUTOMATICS 2013 Vol. 17 No. 1 http://dx.doi.org/10.7494/automat.2013.17.1.59 Rados³aw Klimek* From Workflow Design Patterns to Logical Specifications 1. Introduction Formal methods in software
Lecture 16 : Relations and Functions DRAFT
CS/Math 240: Introduction to Discrete Mathematics 3/29/2011 Lecture 16 : Relations and Functions Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT In Lecture 3, we described a correspondence
Enforcing Data Quality Rules for a Synchronized VM Log Audit Environment Using Transformation Mapping Techniques
Enforcing Data Quality Rules for a Synchronized VM Log Audit Environment Using Transformation Mapping Techniques Sean Thorpe 1, Indrajit Ray 2, and Tyrone Grandison 3 1 Faculty of Engineering and Computing,
AUTOMATED TEST GENERATION FOR SOFTWARE COMPONENTS
TKK Reports in Information and Computer Science Espoo 2009 TKK-ICS-R26 AUTOMATED TEST GENERATION FOR SOFTWARE COMPONENTS Kari Kähkönen ABTEKNILLINEN KORKEAKOULU TEKNISKA HÖGSKOLAN HELSINKI UNIVERSITY OF
Scaling 10Gb/s Clustering at Wire-Speed
Scaling 10Gb/s Clustering at Wire-Speed InfiniBand offers cost-effective wire-speed scaling with deterministic performance Mellanox Technologies Inc. 2900 Stender Way, Santa Clara, CA 95054 Tel: 408-970-3400
Chapter 4 Software Lifecycle and Performance Analysis
Chapter 4 Software Lifecycle and Performance Analysis This chapter is aimed at illustrating performance modeling and analysis issues within the software lifecycle. After having introduced software and
Bounded Treewidth in Knowledge Representation and Reasoning 1
Bounded Treewidth in Knowledge Representation and Reasoning 1 Reinhard Pichler Institut für Informationssysteme Arbeitsbereich DBAI Technische Universität Wien Luminy, October 2010 1 Joint work with G.
Tool Support for Software Variability Management and Product Derivation in Software Product Lines
Tool Support for Software Variability Management and Product Derivation in Software s Hassan Gomaa 1, Michael E. Shin 2 1 Dept. of Information and Software Engineering, George Mason University, Fairfax,
Meta Model Based Integration of Role-Based and Discretionary Access Control Using Path Expressions
Meta Model Based Integration of Role-Based and Discretionary Access Control Using Path Expressions Kathrin Lehmann, Florian Matthes Chair for Software Engineering for Business Information Systems Technische
Checking Access to Protected Members in the Java Virtual Machine
Checking Access to Protected Members in the Java Virtual Machine Alessandro Coglio Kestrel Institute 3260 Hillview Avenue, Palo Alto, CA 94304, USA Ph. +1-650-493-6871 Fax +1-650-424-1807 http://www.kestrel.edu/
Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows
TECHNISCHE UNIVERSITEIT EINDHOVEN Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows Lloyd A. Fasting May 2014 Supervisors: dr. M. Firat dr.ir. M.A.A. Boon J. van Twist MSc. Contents
131-1. Adding New Level in KDD to Make the Web Usage Mining More Efficient. Abstract. 1. Introduction [1]. 1/10
1/10 131-1 Adding New Level in KDD to Make the Web Usage Mining More Efficient Mohammad Ala a AL_Hamami PHD Student, Lecturer m_ah_1@yahoocom Soukaena Hassan Hashem PHD Student, Lecturer soukaena_hassan@yahoocom
Visualizing Information Flow through C Programs
Visualizing Information Flow through C Programs Joe Hurd, Aaron Tomb and David Burke Galois, Inc. {joe,atomb,davidb}@galois.com Systems Software Verification Workshop 7 October 2010 Joe Hurd, Aaron Tomb
estatistik.core: COLLECTING RAW DATA FROM ERP SYSTEMS
WP. 2 ENGLISH ONLY UNITED NATIONS STATISTICAL COMMISSION and ECONOMIC COMMISSION FOR EUROPE CONFERENCE OF EUROPEAN STATISTICIANS Work Session on Statistical Data Editing (Bonn, Germany, 25-27 September
Clarifying a vision on certification of MDA tools
SCIENTIFIC PAPERS, UNIVERSITY OF LATVIA, 2010. Vol. 757 COMPUTER SCIENCE AND INFORMATION TECHNOLOGIES 23 29 P. Clarifying a vision on certification of MDA tools Antons Cernickins Riga Technical University,
A Configuration Management Model for Software Product Line
A Configuration Management Model for Software Product Line Liguo Yu 1 and Srini Ramaswamy 2 1 Computer Science and Informatics Indiana University South Bend South Bend, IN 46634, USA [email protected] 2 Computer
Project and Production Management Prof. Arun Kanda Department of Mechanical Engineering Indian Institute of Technology, Delhi
Project and Production Management Prof. Arun Kanda Department of Mechanical Engineering Indian Institute of Technology, Delhi Lecture - 9 Basic Scheduling with A-O-A Networks Today we are going to be talking
A first step towards modeling semistructured data in hybrid multimodal logic
A first step towards modeling semistructured data in hybrid multimodal logic Nicole Bidoit * Serenella Cerrito ** Virginie Thion * * LRI UMR CNRS 8623, Université Paris 11, Centre d Orsay. ** LaMI UMR
Firewall Verification and Redundancy Checking are Equivalent
Firewall Verification and Redundancy Checking are Equivalent H. B. Acharya University of Texas at Austin [email protected] M. G. Gouda National Science Foundation University of Texas at Austin [email protected]
Ruby on Rails Tutorial - Bounded Verification of Data Models
Bounded Verification of Ruby on Rails Data Models Jaideep Nijjar and Tevfik Bultan University of California, Santa Barbara {jaideepnijjar, [email protected] ABSTRACT The use of scripting languages to
A Serial Partitioning Approach to Scaling Graph-Based Knowledge Discovery
A Serial Partitioning Approach to Scaling Graph-Based Knowledge Discovery Runu Rathi, Diane J. Cook, Lawrence B. Holder Department of Computer Science and Engineering The University of Texas at Arlington
Object-Oriented Software Specification in Programming Language Design and Implementation
Object-Oriented Software Specification in Programming Language Design and Implementation Barrett R. Bryant and Viswanathan Vaidyanathan Department of Computer and Information Sciences University of Alabama
Investment Analysis using the Portfolio Analysis Machine (PALMA 1 ) Tool by Richard A. Moynihan 21 July 2005
Investment Analysis using the Portfolio Analysis Machine (PALMA 1 ) Tool by Richard A. Moynihan 21 July 2005 Government Investment Analysis Guidance Current Government acquisition guidelines mandate the
Karunya University Dept. of Information Technology
PART A Questions 1. Mention any two software process models. 2. Define risk management. 3. What is a module? 4. What do you mean by requirement process? 5. Define integration testing. 6. State the main
A Business Process Services Portal
A Business Process Services Portal IBM Research Report RZ 3782 Cédric Favre 1, Zohar Feldman 3, Beat Gfeller 1, Thomas Gschwind 1, Jana Koehler 1, Jochen M. Küster 1, Oleksandr Maistrenko 1, Alexandru
Automated Theorem Proving - summary of lecture 1
Automated Theorem Proving - summary of lecture 1 1 Introduction Automated Theorem Proving (ATP) deals with the development of computer programs that show that some statement is a logical consequence of
Query-aware Test Generation Using a Relational Constraint Solver
Query-aware Test Generation Using a Relational Constraint Solver Shadi Abdul Khalek Bassem Elkarablieh Yai O. Laleye Sarfraz Khurshid The University of Texas at Austin {sabdulkhalek, elkarabl, lalaye,
Software Engineering Reference Framework
Software Engineering Reference Framework Michel Chaudron, Jan Friso Groote, Kees van Hee, Kees Hemerik, Lou Somers, Tom Verhoeff. Department of Mathematics and Computer Science Eindhoven University of
Abstraction in Computer Science & Software Engineering: A Pedagogical Perspective
Orit Hazzan's Column Abstraction in Computer Science & Software Engineering: A Pedagogical Perspective This column is coauthored with Jeff Kramer, Department of Computing, Imperial College, London ABSTRACT
A Comparison of General Approaches to Multiprocessor Scheduling
A Comparison of General Approaches to Multiprocessor Scheduling Jing-Chiou Liou AT&T Laboratories Middletown, NJ 0778, USA [email protected] Michael A. Palis Department of Computer Science Rutgers University
A Framework of Model-Driven Web Application Testing
A Framework of Model-Driven Web Application Testing Nuo Li, Qin-qin Ma, Ji Wu, Mao-zhong Jin, Chao Liu Software Engineering Institute, School of Computer Science and Engineering, Beihang University, China
Interactive Graphic Design Using Automatic Presentation Knowledge
Interactive Graphic Design Using Automatic Presentation Knowledge Steven F. Roth, John Kolojejchick, Joe Mattis, Jade Goldstein School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213
Integrating Pattern Mining in Relational Databases
Integrating Pattern Mining in Relational Databases Toon Calders, Bart Goethals, and Adriana Prado University of Antwerp, Belgium {toon.calders, bart.goethals, adriana.prado}@ua.ac.be Abstract. Almost a
An eclipse-based Feature Models toolchain
An eclipse-based Feature Models toolchain Luca Gherardi, Davide Brugali Dept. of Information Technology and Mathematics Methods, University of Bergamo [email protected], [email protected] Abstract.
Determination of the normalization level of database schemas through equivalence classes of attributes
Computer Science Journal of Moldova, vol.17, no.2(50), 2009 Determination of the normalization level of database schemas through equivalence classes of attributes Cotelea Vitalie Abstract In this paper,
Modelling and Verification of Business Processes
Modelling and Verification of Business Processes Costin Bădică Department of Computer Science King s College London, WC2R 2LS, UK [email protected] Chris Fox Department of Computer Science University
Simulating the Structural Evolution of Software
Simulating the Structural Evolution of Software Benjamin Stopford 1, Steve Counsell 2 1 School of Computer Science and Information Systems, Birkbeck, University of London 2 School of Information Systems,
Cost Model: Work, Span and Parallelism. 1 The RAM model for sequential computation:
CSE341T 08/31/2015 Lecture 3 Cost Model: Work, Span and Parallelism In this lecture, we will look at how one analyze a parallel program written using Cilk Plus. When we analyze the cost of an algorithm
Protein Protein Interaction Networks
Functional Pattern Mining from Genome Scale Protein Protein Interaction Networks Young-Rae Cho, Ph.D. Assistant Professor Department of Computer Science Baylor University it My Definition of Bioinformatics
UML-based Test Generation and Execution
UML-based Test Generation and Execution Jean Hartmann, Marlon Vieira, Herb Foster, Axel Ruder Siemens Corporate Research, Inc. 755 College Road East Princeton NJ 08540, USA [email protected] ABSTRACT
Co-Creation of Models and Metamodels for Enterprise. Architecture Projects.
Co-Creation of Models and Metamodels for Enterprise Architecture Projects Paola Gómez [email protected] Hector Florez [email protected] ABSTRACT The linguistic conformance and the ontological
Umbrella: A New Component-Based Software Development Model
2009 International Conference on Computer Engineering and Applications IPCSIT vol.2 (2011) (2011) IACSIT Press, Singapore Umbrella: A New Component-Based Software Development Model Anurag Dixit and P.C.
Mathematics for Computer Science/Software Engineering. Notes for the course MSM1F3 Dr. R. A. Wilson
Mathematics for Computer Science/Software Engineering Notes for the course MSM1F3 Dr. R. A. Wilson October 1996 Chapter 1 Logic Lecture no. 1. We introduce the concept of a proposition, which is a statement
CHAPTER 1 INTRODUCTION
1 CHAPTER 1 INTRODUCTION Exploration is a process of discovery. In the database exploration process, an analyst executes a sequence of transformations over a collection of data structures to discover useful
Sudoku as a SAT Problem
Sudoku as a SAT Problem Inês Lynce IST/INESC-ID, Technical University of Lisbon, Portugal [email protected] Joël Ouaknine Oxford University Computing Laboratory, UK [email protected] Abstract Sudoku
Semantic Search in Portals using Ontologies
Semantic Search in Portals using Ontologies Wallace Anacleto Pinheiro Ana Maria de C. Moura Military Institute of Engineering - IME/RJ Department of Computer Engineering - Rio de Janeiro - Brazil [awallace,anamoura]@de9.ime.eb.br
Technology WHITE PAPER
Technology WHITE PAPER What We Do Neota Logic builds software with which the knowledge of experts can be delivered in an operationally useful form as applications embedded in business systems or consulted
Lightweight Service-Based Software Architecture
Lightweight Service-Based Software Architecture Mikko Polojärvi and Jukka Riekki Intelligent Systems Group and Infotech Oulu University of Oulu, Oulu, Finland {mikko.polojarvi,jukka.riekki}@ee.oulu.fi
SOFT 437. Software Performance Analysis. Ch 5:Web Applications and Other Distributed Systems
SOFT 437 Software Performance Analysis Ch 5:Web Applications and Other Distributed Systems Outline Overview of Web applications, distributed object technologies, and the important considerations for SPE
VisCG: Creating an Eclipse Call Graph Visualization Plug-in. Kenta Hasui, Undergraduate Student at Vassar College Class of 2015
VisCG: Creating an Eclipse Call Graph Visualization Plug-in Kenta Hasui, Undergraduate Student at Vassar College Class of 2015 Abstract Call graphs are a useful tool for understanding software; however,
A Model-driven Approach to Predictive Non Functional Analysis of Component-based Systems
A Model-driven Approach to Predictive Non Functional Analysis of Component-based Systems Vincenzo Grassi Università di Roma Tor Vergata, Italy Raffaela Mirandola {vgrassi, mirandola}@info.uniroma2.it Abstract.
PLEDGE: A Product Line Editor and Test Generation Tool
PLEDGE: A Product Line Editor and Test Generation Tool Christopher Henard [email protected] Jacques Klein [email protected] Mike Papadakis [email protected] Yves Le Traon [email protected]
A Framework for the Delivery of Personalized Adaptive Content
A Framework for the Delivery of Personalized Adaptive Content Colm Howlin CCKF Limited Dublin, Ireland [email protected] Danny Lynch CCKF Limited Dublin, Ireland [email protected] Abstract
University of Potsdam Faculty of Computer Science. Clause Learning in SAT Seminar Automatic Problem Solving WS 2005/06
University of Potsdam Faculty of Computer Science Clause Learning in SAT Seminar Automatic Problem Solving WS 2005/06 Authors: Richard Tichy, Thomas Glase Date: 25th April 2006 Contents 1 Introduction
Object Oriented Programming. Risk Management
Section V: Object Oriented Programming Risk Management In theory, there is no difference between theory and practice. But, in practice, there is. - Jan van de Snepscheut 427 Chapter 21: Unified Modeling
How To Know If A Domain Is Unique In An Octempo (Euclidean) Or Not (Ecl)
Subsets of Euclidean domains possessing a unique division algorithm Andrew D. Lewis 2009/03/16 Abstract Subsets of a Euclidean domain are characterised with the following objectives: (1) ensuring uniqueness
Software Verification: Infinite-State Model Checking and Static Program
Software Verification: Infinite-State Model Checking and Static Program Analysis Dagstuhl Seminar 06081 February 19 24, 2006 Parosh Abdulla 1, Ahmed Bouajjani 2, and Markus Müller-Olm 3 1 Uppsala Universitet,
A Framework of Personalized Intelligent Document and Information Management System
A Framework of Personalized Intelligent and Information Management System Xien Fan Department of Computer Science, College of Staten Island, City University of New York, Staten Island, NY 10314, USA Fang
2. MOTIVATING SCENARIOS 1. INTRODUCTION
Multiple Dimensions of Concern in Software Testing Stanley M. Sutton, Jr. EC Cubed, Inc. 15 River Road, Suite 310 Wilton, Connecticut 06897 [email protected] 1. INTRODUCTION Software testing is an area
The Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
Dsc+Mock: A Test Case + Mock Class Generator in Support of Coding Against Interfaces
Dsc+Mock: A Test Case + Mock Class Generator in Support of Coding Against Interfaces Mainul Islam, Christoph Csallner Computer Science and Engineering Department University of Texas at Arlington Arlington,
To introduce software process models To describe three generic process models and when they may be used
Software Processes Objectives To introduce software process models To describe three generic process models and when they may be used To describe outline process models for requirements engineering, software
EVALUATION. WA1844 WebSphere Process Server 7.0 Programming Using WebSphere Integration COPY. Developer
WA1844 WebSphere Process Server 7.0 Programming Using WebSphere Integration Developer Web Age Solutions Inc. USA: 1-877-517-6540 Canada: 1-866-206-4644 Web: http://www.webagesolutions.com Chapter 6 - Introduction
Modeling Guidelines Manual
Modeling Guidelines Manual [Insert company name here] July 2014 Author: John Doe [email protected] Page 1 of 22 Table of Contents 1. Introduction... 3 2. Business Process Management (BPM)... 4 2.1.
Lecture 17 : Equivalence and Order Relations DRAFT
CS/Math 240: Introduction to Discrete Mathematics 3/31/2011 Lecture 17 : Equivalence and Order Relations Instructor: Dieter van Melkebeek Scribe: Dalibor Zelený DRAFT Last lecture we introduced the notion
Introduction to Formal Methods. Các Phương Pháp Hình Thức Cho Phát Triển Phần Mềm
Introduction to Formal Methods Các Phương Pháp Hình Thức Cho Phát Triển Phần Mềm Outline Introduction Formal Specification Formal Verification Model Checking Theorem Proving Introduction Good papers to
Ontology-Based Discovery of Workflow Activity Patterns
Ontology-Based Discovery of Workflow Activity Patterns Diogo R. Ferreira 1, Susana Alves 1, Lucinéia H. Thom 2 1 IST Technical University of Lisbon, Portugal {diogo.ferreira,susana.alves}@ist.utl.pt 2
(Refer Slide Time: 01:52)
Software Engineering Prof. N. L. Sarda Computer Science & Engineering Indian Institute of Technology, Bombay Lecture - 2 Introduction to Software Engineering Challenges, Process Models etc (Part 2) This
Object Oriented Design
Object Oriented Design Kenneth M. Anderson Lecture 20 CSCI 5828: Foundations of Software Engineering OO Design 1 Object-Oriented Design Traditional procedural systems separate data and procedures, and
Chapter 7 Application Protocol Reference Architecture
Application Protocol Reference Architecture Chapter 7 Application Protocol Reference Architecture This chapter proposes an alternative reference architecture for application protocols. The proposed reference
Requirements Analysis Concepts & Principles. Instructor: Dr. Jerry Gao
Requirements Analysis Concepts & Principles Instructor: Dr. Jerry Gao Requirements Analysis Concepts and Principles - Requirements Analysis - Communication Techniques - Initiating the Process - Facilitated
ONTOLOGY FOR MOBILE PHONE OPERATING SYSTEMS
ONTOLOGY FOR MOBILE PHONE OPERATING SYSTEMS Hasni Neji and Ridha Bouallegue Innov COM Lab, Higher School of Communications of Tunis, Sup Com University of Carthage, Tunis, Tunisia. Email: [email protected];
SQLMutation: A tool to generate mutants of SQL database queries
SQLMutation: A tool to generate mutants of SQL database queries Javier Tuya, Mª José Suárez-Cabal, Claudio de la Riva University of Oviedo (SPAIN) {tuya cabal claudio} @ uniovi.es Abstract We present a
Schema Mappings and Data Exchange
Schema Mappings and Data Exchange Phokion G. Kolaitis University of California, Santa Cruz & IBM Research-Almaden EASSLC 2012 Southwest University August 2012 1 Logic and Databases Extensive interaction
Binary search tree with SIMD bandwidth optimization using SSE
Binary search tree with SIMD bandwidth optimization using SSE Bowen Zhang, Xinwei Li 1.ABSTRACT In-memory tree structured index search is a fundamental database operation. Modern processors provide tremendous
Exponential time algorithms for graph coloring
Exponential time algorithms for graph coloring Uriel Feige Lecture notes, March 14, 2011 1 Introduction Let [n] denote the set {1,..., k}. A k-labeling of vertices of a graph G(V, E) is a function V [k].
