From Blackbox Fuzzing to Whitebox Fuzzing towards Verification
|
|
|
- Elfreda Nicholson
- 10 years ago
- Views:
Transcription
1 From Blackbox Fuzzing to Whitebox Fuzzing towards Verification Blackbox Fuzzing Whitebox Fuzzing Verification Patrice Godefroid Microsoft Research Page 1 July 2010
2 Acknowledgments Joint work with: MSR: Ella Bounimova, David Molnar, CSE: Michael Levin, Chris Marsh, Lei Fang, Stuart de Jong, Interns Dennis Jeffries (06), David Molnar (07), Adam Kiezun (07), Bassem Elkarablieh (08), Cindy Rubio-Gonzalez (08,09), Johannes Kinder (09), Thanks to the entire SAGE team and users! Z3: Nikolaj Bjorner, Leonardo de Moura, Windows: Nick Bartmon, Eric Douglas, Office: Tom Gallagher, Eric Jarvi, Octavian Timofte, SAGE users all across Microsoft! Page 2 July 2010
3 References See DART: Directed Automated Random Testing, with Klarlund and Sen, PLDI 2005 Compositional Dynamic Test Generation, POPL 2007 Automated Whitebox Fuzz Testing, with Levin and Molnar, NDSS 2008 Demand-Driven Compositional Symbolic Execution, with Anand and Tillmann, TACAS 2008 Grammar-Based Whitebox Fuzzing, with Kiezun and Levin, PLDI 2008 Active Property Checking, with Levin and Molnar, EMSOFT 2008 Precise Pointer Reasoning for Dynamic Test Generation, with Elkarablieh and Levin, ISSTA 2009 Compositional May-Must Program Analysis: Unleashing The Power of Alternation, with Nori, Rajamani and Tetali, POPL 2010 Proving Memory Safety of Floating-Point Computations by Combining Static and Dynamic Program Analysis, with Kinder, Page 3 July 2010
4 Security is Critical (to Microsoft) Software security bugs can be very expensive: Cost of each Microsoft Security Bulletin: $Millions Cost due to worms (Slammer, CodeRed, Blaster, etc.): $Billions Many security exploits are initiated via files or packets Ex: MS Windows includes parsers for hundreds of file formats Security testing: hunting for million-dollar bugs Write A/V (always exploitable), Read A/V (sometimes exploitable), NULL-pointer dereference, division-by-zero (harder to exploit but still DOS attacks), etc. Page 4 July 2010
5 Hunting for Security Bugs I am from Belgium too! Main techniques used by black hats : Code inspection (of binaries) and Blackbox fuzz testing Blackbox fuzz testing: A form of blackbox random testing [Miller+90] Randomly fuzz (=modify) a well-formed input Grammar-based fuzzing: rules that encode well-formed ness + heuristics about how to fuzz (e.g., using probabilistic weights) Heavily used in security testing Simple yet effective: many bugs found this way At Microsoft, fuzzing is mandated by the SDL Page 5 July 2010
6 Blackbox Fuzzing Examples: Peach, Protos, Spike, Autodafe, etc. Why so many blackbox fuzzers? Because anyone can write (a simple) one in a week-end! Conceptually simple, yet effective Sophistication is in the add-on Test harnesses (e.g., for packet fuzzing) Grammars (for specific input formats) Note: usually, no principled spec-based test generation No attempt to cover each state/rule in the grammar When probabilities, no global optimization (simply random walks) Page 6 July 2010
7 Introducing Whitebox Fuzzing Idea: mix fuzz testing with dynamic test generation Symbolic execution Collect constraints on inputs Negate those, solve with constraint solver, generate new inputs do systematic dynamic test generation (=DART) Whitebox Fuzzing = DART meets Fuzz Two Parts: 1. Foundation: DART (Directed Automated Random Testing) 2. Key extensions ( Whitebox Fuzzing ), implemented in SAGE Page 7 July 2010
8 Automatic Code-Driven Test Generation Problem: Given a sequential program with a set of input parameters, generate a set of inputs that maximizes code coverage = automate test generation using program analysis This is not model-based testing (= generate tests from an FSM spec) Page 8 July 2010
9 How? (1) Static Test Generation Static analysis to partition the program s input space [King76, ] Ineffective whenever symbolic reasoning is not possible which is frequent in practice (pointer manipulations, complex arithmetic, calls to complex OS or library functions, etc.) Example: int obscure(int x, int y) { } if (x==hash(y)) error(); return 0; Can t statically generate values for x and y that satisfy x==hash(y)! Page 9 July 2010
10 How? (2) Dynamic Test Generation Run the program (starting with some random inputs), gather constraints on inputs at conditional statements, use a constraint solver to generate new test inputs Repeat until a specific program statement is reached [Korel90, ] Or repeat to try to cover ALL feasible program paths: DART = Directed Automated Random Testing = systematic dynamic test generation [PLDI 05, ] detect crashes, assertion violations, use runtime checkers (Purify, ) Page 10 July 2010
11 DART = Directed Automated Random Testing Example: Run 1 : - start with (random) x=33, y=42 int obscure(int x, int y) { - execute concretely and symbolically: if (33!= 567) if (x!= hash(y)) if (x==hash(y)) error(); constraint too complex return 0; simplify it: x!= 567 } - solve: x==567 solution: x=567 - new test input: x=567, y=42 Observations: Run 2 : the other branch is executed All program paths are now covered! Dynamic test generation extends static test generation with additional runtime information: it is more powerful The number of program paths can be infinite: may not terminate! Still, DART works well for small programs (1,000s LOC) Significantly improves code coverage vs. random testing Page 11 July 2010
12 DART Implementations Defined by symbolic execution, constraint generation and solving Languages: C, Java, x86,.net, Theories: linear arith., bit-vectors, arrays, uninterpreted functions, Solvers: lp_solve, CVCLite, STP, Disolver, Z3, Examples of tools/systems implementing DART: EXE/EGT (Stanford): independent [ 05-06] closely related work CUTE = same as first DART implementation done at Bell Labs SAGE (CSE/MSR) for x86 binaries and merges it with fuzz testing for finding security bugs (more later) PEX (MSR) for.net binaries in conjunction with parameterized-unit tests for unit testing of.net programs YOGI (MSR) for checking the feasibility of program paths generated statically using a SLAM-like tool Vigilante (MSR) for generating worm filters BitScope (CMU/Berkeley) for malware analysis CatchConv (Berkeley) focus on integer overflows Splat (UCLA) focus on fast detection of buffer overflows Apollo (MIT) for testing web applications and more! Page 12 July 2010
13 Whitebox Fuzzing [NDSS 08] Whitebox Fuzzing = DART meets Fuzz Apply DART to large applications (not unit) Start with a well-formed input (not random) Combine with a generational search (not DFS) Negate 1-by-1 each constraint in a path constraint Generate many children for each parent run Challenge all the layers of the application sooner Leverage expensive symbolic execution Search spaces are huge, the search is partial yet effective at finding bugs! parent Gen 1 Page 13 July 2010
14 Example void top(char input[4]) input = good { int cnt = 0; Path constraint: if (input[0] == b ) cnt++; I 0!= b I 0 = b bood if (input[1] == a ) cnt++; I 1!= a I 1 = a gaod if (input[2] == d ) cnt++; I 2!= d I 2 = d godd if (input[3] ==! ) cnt++; I 3!=! I 3 =! goo! } if (cnt >= 3) crash(); Negate each constraint in path constraint Solve new constraint new input good Gen 1 Page 14 July 2010
15 The Search Space void top(char input[4]) { int cnt = 0; if (input[0] == b ) cnt++; if (input[1] == a ) cnt++; if (input[2] == d ) cnt++; if (input[3] ==! ) cnt++; if (cnt >= 3) crash(); } Page 15 July 2010
16 SAGE (Scalable Automated Guided Execution) Generational search introduced in SAGE Performs symbolic execution of x86 execution traces Builds on Nirvana, idna and TruScan for x86 analysis Don t care about language or build process Easy to test new applications, no interference possible Can analyse any file-reading Windows applications Several optimizations to handle huge execution traces Constraint caching and common subexpression elimination Unrelated constraint optimization Constraint subsumption for constraints from input-bound loops Flip-count limit (to prevent endless loop expansions) Page 16 July 2010
17 SAGE Architecture Input0 Coverage Data Constraints Check for Crashes (AppVerifier) Code Coverage (Nirvana) Generate Constraints (TruScan) Solve Constraints (Z3) Input1 Input2 SAGE was mostly developed by CSE InputN MSR algorithms & code inside Page 17 July 2010
18 Some Experiments Most much (100x) bigger than ever tried before! Seven applications 10 hours search each App Tested #Tests Mean Depth Mean #Instr. Mean Input Size ANI ,066,087 5,400 Media ,409,376 65,536 Media ,432,489 27,335 Media ,644,652 30,833 Media ,685,240 22,209 Compressed File Format , OfficeApp ,731,248 45,064 Page 18 July 2010
19 Generational Search Leverages Symbolic Execution Each symbolic execution is expensive m30s SymbolicExecutor TestTask X1,000 Yet, symbolic execution does not dominate search time SymbolicExecutor 10 hours Testing/Tracing/Coverage Page 19 July 2010
20 SAGE Results Since April 07 1 st release: many new security bugs found (missed by blackbox fuzzers, static analysis) Apps: image processors, media players, file decoders, Bugs: Write A/Vs, Read A/Vs, Crashes, Many triaged as security critical, severity 1, priority 1 (would trigger Microsoft security bulletin if known outside MS) Example: WEX Security team for Win7 Dedicated fuzzing lab with 100s machines 100s apps (deployed on 1billion+ computers) ~1/3 of all fuzzing bugs found by SAGE! SAGE = gold medal at Fuzzing Olympics organized by SWI at BlueHat 08 (Oct 08) Credit due to entire SAGE team + users! Page 20 July 2010
21 WEX Fuzz Dashboard Snippet Page 21 July 2010
22 WEX Fuzzing Lab Bug Yield for Win7 How fuzzing bugs found ( ) : 100s of apps, total number of fuzzing bugs is confidential Default Blackbox Fuzzer + Regression All Others SAGE SAGE is running 24/7 on 100s machines: the largest usage ever of any SMT solver N. Bjorner + L. de Moura (MSR, Z3 authors) But SAGE didn t exist in 2006 Since 2007 (SAGE 1 st release), ~1/3 bugs found by SAGE But SAGE currently deployed on only ~2/3 of those apps Normalizing the data by 2/3, SAGE found ~1/2 bugs SAGE is more CPU expensive, so it is run last in the lab, so all SAGE bugs were missed by everything else! Page 22 July 2010
23 SAGE Summary SAGE is so effective at finding bugs that, for the first time, we face bug triage issues with dynamic test generation What makes it so effective? Works on large applications (not unit test, like DART, EXE, etc.) Can detect bugs due to problems across components Fully automated (focus on file fuzzing) Easy to deploy (x86 analysis any language or build process!) 1 st tool for whole-program dynamic symbolic execution at x86 level Now, used daily in various groups at Microsoft Page 23 July 2010
24 More On the Research Behind SAGE How to recover from imprecision in symbolic execution? PLDI 05 Must under-approximations How to scale symbolic exec. to billions of instructions? NDSS 08 Techniques to deal with large path constraints How to check efficiently many properties together? EMSOFT 08 Active property checking How to leverage grammars for complex input formats? PLDI 08 Lift input constraints to the level of symbolic terminals in an input grammar How to deal with path explosion? POPL 07, TACAS 08, POPL 10 Symbolic test summaries (more later) How to reason precisely about pointers? ISSTA 09 New memory models leveraging concrete memory addresses and regions How to deal with floating-point instructions? ISSTA 10 Prove non-interference with memory accesses + research on constraint solvers (Z3, disolver,...) Page 24 July 2010
25 What Next? Towards Verification When can we safely stop testing? When we know that there are no more bugs! = Verification Testing can only prove the existence of bugs, not their absence. Unless it is exhaustive! This is the model checking thesis Model Checking = exhaustive testing (state-space exploration) Two main approaches to software model checking: Modeling languages state-space exploration Model checking [Dijkstra] abstraction (SLAM, Bandera, FeaVer, BLAST, ) adaptation Programming languages state-space exploration Systematic testing Concurrency: VeriSoft, JPF, CMC, Bogor, CHESS, Data inputs: DART, EXE, SAGE, Page 25 July 2010
26 Exhaustive Testing? Model checking is always up to some bound Limited (often finite) input domain, for specific properties, under some environment assumptions Ex: exhaustive testing of Win7 JPEG parser up to 1,000 input bytes 8000 bits 2^8000 possibilities if 1 test per sec, 2^8000 secs FYI, 15 billion years = secs = 2^60 secs! MUST be symbolic! How far can we go? Practical goals: (easier?) Eradicate all remaining buffer overflows in all Windows parsers Better coverage guarantees to justify no new bug found Reduce costs & risks for Microsoft: when to stop fuzzing? Increase costs & risks for Black Hats! Many have probably moved to greener pastures already (Ex: Adobe) Ex: <5 security bulletins in all the SAGE-cleaned Win7 parsers If noone can find bugs in P, P is observationally equivalent to verified! Page 26 July 2010
27 How to Get There? 1. Identify and patch holes in symbolic execution + constraint solving 2. Tackle path explosion with compositional testing and symbolic test summaries [POPL 07,TACAS 08,POPL 10] Fuzzing in the (Virtual) Cloud (Sagan) New centralized server collecting stats from all SAGE runs! Track results (bugs, concrete & symbolic test coverage), incompleteness (unhandled tainted x86 instructions, Z3 timeouts, divergences, etc.) Help troubleshooting (SAGE has 100+ options ) Tell us what works and what does not Page 27 July 2010
28 The Art of Constraint Generation Static analysis: abstract away irrelevant details Good for focused search, can be combined with DART (Ex: [POPL 10]) But for bit-precise analysis of low-level code (function pointers, in-lined assembly,...)? In a non-property-guided setting? Open problem Bit-precise VC-gen: statically generate 1 formula from a program Good to prove complex properties of small programs (units) Does not scale (huge formula encodings), asks too much of the user SAT/SMT-based Bounded Model Checking : stripped-down VC-gen Emphasis on automation Unrolling all loops is naïve, does not scale DART : the only option today for large programs (Ex: Excel) Path-by-path exploration is naïve, but whitebox fuzzing can scale it to large executions (Z3 is not the bottleneck) + zero false alarms! But suffers from path explosion Page 28 July 2010
29 DART is Beautiful Generates formulas where the only free symbolic variables are whole-program inputs When generating tests, one can only control inputs! Strength: scalability to large programs Only tracks direct input dependencies (i.e., tests on inputs); the rest of the execution is handled with the best constantpropagation engine ever: running the code on the computer! (The size of) path constraints only depend on (the number of) program tests on inputs, not on the size of the program = the right metric: complexity only depends on nondeterminism! Price to pay: path explosion [POPL 07] Solution = symbolic test summaries Page 29 July 2010
30 Example void top(char input[4]) { int cnt = 0; input = good Path constraint: if (input[0] == b ) cnt++; if (input[1] == a ) cnt++; if (input[2] == d ) cnt++; if (input[3] ==! ) cnt++; I 0!= b I 1!= a I 2!= d I 3!=! if (cnt >= 3) crash(); } Page 30 July 2010
31 Compositionality = Key to Scalability Idea: compositional dynamic test generation [POPL 07] use summaries of individual functions (or program blocks, etc.) like in interprocedural static analysis but here must formulas generated dynamically If f calls g, test g, summarize the results, and use g s summary when testing f A summary φ(g) is a disjunction of path constraints expressed in terms of g s input preconditions and g s output postconditions: φ(g) = φ(w) with φ(w) = pre(w) post(w) g s outputs are treated as fresh symbolic inputs to f, all bound to prior inputs and can be eliminated (for test generation) Can provide same path coverage exponentially faster! See details and refinements in [POPL 07,TACAS 08,POPL 10] Page 31 July 2010
32 The Engineering of Test Summaries Systematically summarizing everywhere is foolish Very expensive and not necessary (costs outweigh benefits) Don t fall into the VC-gen or BMC traps! Summarization on-demand: (100% algorithmic) When? At search bottlenecks (with dynamic feedback loop) Where? At simple interfaces (with simple data types) How? With limited side-effects (to be manageable and sound ) Goal: use summaries intelligently THE KEY to scalable bit-precise whole-program analysis? It is necessary! But in what form(s)? Is it sufficient? Stay tuned Page 32 July 2010
33 Summaries Cure Search Redundancy Across different program paths IF THEN ELSE Across different program versions [ Incremental Compositional Dynamic Test Generation, with S. Lahiri and C. Rubio-Gonzalez, MSR TR, Feb 2010] Across different applications Summaries avoid unnecessary work What if central server of summaries for all code?... Sagan 2.0 Page 33 July 2010
34 Conclusion: Impact of SAGE (In Numbers) 100+ machine-years Runs in the largest dedicated fuzzing lab in the world 100+ million constraints Largest computational usage ever for any SMT solver 100s of apps, 100s of bugs (missed by everything else) Bug fixes shipped to 1 Billion+ computers worldwide Millions of dollars saved for Microsoft + time/energy savings for the world DART, Whitebox fuzzing now adopted by (many) others (10s tools, 100s citations) Page 34 July 2010
35 Conclusion: Blackbox vs. Whitebox Fuzzing Different cost/precision tradeoffs Blackbox is lightweight, easy and fast, but poor coverage Whitebox is smarter, but complex and slower Note: other recent semi-whitebox approaches Less smart (no symbolic exec, constr. solving) but more lightweight: Flayer (taint-flow, may generate false alarms), Bunny-the-fuzzer (taint-flow, source-based, fuzz heuristics from input usage), etc. Which is more effective at finding bugs? It depends Many apps are so buggy, any form of fuzzing find bugs in those! Once low-hanging bugs are gone, fuzzing must become smarter: use whitebox and/or user-provided guidance (grammars, etc.) Bottom-line: in practice, use both! (We do at Microsoft) Page 35 July 2010
36 Myth: Blackbox is Cheaper than Whitebox Wrong! A whitebox fuzzer is slower than a blackbox fuzzer Whitebox needs more time to generate new tests (since smarter) But over time, blackbox is more expensive than whitebox Blackbox does not know when to (soundly) stop! Example: int magic(int input){ } if (input==645723) error(); return 0; coverage (bugs) whitebox stop! blackbox time (money) Page 36 July 2010
37 What Next? Towards Verification Tracking all(?) sources of incompleteness Summaries (on-demand ) against path explosion How far can we go? Reduce costs & risks for Microsoft: when to stop fuzzing? Increase costs & risks for Black Hats (goal already achieved?) For history books: Blackbox Fuzzing Whitebox Fuzzing Verification There is one thing stronger than all the armies in the world; and that is an idea whose time has come. -- Victor Hugo Page 37 July 2010
38 What Next? Towards Verification Tracking all(?) sources of incompleteness Summaries (on-demand ) against path explosion How far can we go? Reduce costs & risks for Microsoft: when to stop fuzzing? Increase costs & risks for Blackhats (goal already achieved?) For history books: Blackbox Fuzzing Whitebox Fuzzing Verification There is one thing stronger than all the armies in the world; and that is an idea whose time has come. -- Victor Hugo Page 38 July 2010
Software Model Checking Improving Security of a Billion Computers
Software Model Checking Improving Security of a Billion Computers Patrice Godefroid Microsoft Research Page 1 November 2009 Acknowledgments Joint work with Michael Levin (CSE) and others: Chris Marsh,
Applications of SMT to. Test Generation
Applications of SMT to Test Generation Patrice Godefroid Microsoft Research Page 1 June 2012 Test Generation is Big Business #1 application for SMT solvers today (CPU usage) SAGE @ Microsoft: 1 st whitebox
Whitebox Fuzzing for Security Testing
Article development led by queue.acm.org SAGE has had a remarkable impact at Microsoft. SAGE: Whitebox Fuzzing for Security Testing doi:10.1145/2093548.2093564 by Patrice Godefroid, Michael Y. Levin, and
SAGE: Whitebox Fuzzing for Security Testing
SAGE: Whitebox Fuzzing for Security Testing SAGE has had a remarkable impact at Microsoft. Patrice Godefroid, Michael Y. Levin, David Molnar, Microsoft Most ACM Queue readers might think of program verification
Dynamic Test Generation for Large Binary Programs
Dynamic Test Generation for Large Binary Programs David Alexander Molnar Electrical Engineering and Computer Sciences University of California at Berkeley Technical Report No. UCB/EECS-2009-157 http://www.eecs.berkeley.edu/pubs/techrpts/2009/eecs-2009-157.html
Model Checking: An Introduction
Announcements Model Checking: An Introduction Meeting 2 Office hours M 1:30pm-2:30pm W 5:30pm-6:30pm (after class) and by appointment ECOT 621 Moodle problems? Fundamentals of Programming Languages CSCI
InvGen: An Efficient Invariant Generator
InvGen: An Efficient Invariant Generator Ashutosh Gupta and Andrey Rybalchenko Max Planck Institute for Software Systems (MPI-SWS) Abstract. In this paper we present InvGen, an automatic linear arithmetic
Towards practical reactive security audit using extended static checkers 1
Towards practical reactive security audit using extended static checkers 1 Julien Vanegue 1 Shuvendu K. Lahiri 2 1 Bloomberg LP, New York 2 Microsoft Research, Redmond May 20, 2013 1 The work was conducted
A Static Analyzer for Large Safety-Critical Software. Considered Programs and Semantics. Automatic Program Verification by Abstract Interpretation
PLDI 03 A Static Analyzer for Large Safety-Critical Software B. Blanchet, P. Cousot, R. Cousot, J. Feret L. Mauborgne, A. Miné, D. Monniaux,. Rival CNRS École normale supérieure École polytechnique Paris
Model Checking of Software
Model Checking of Software Patrice Godefroid Bell Laboratories, Lucent Technologies SpecNCheck Page 1 August 2001 A Brief History of Model Checking Prehistory: transformational programs and theorem proving
Environment Modeling for Automated Testing of Cloud Applications
Environment Modeling for Automated Testing of Cloud Applications Linghao Zhang, Tao Xie, Nikolai Tillmann, Peli de Halleux, Xiaoxing Ma, Jian Lv {lzhang25, txie}@ncsu.edu, {nikolait, jhalleux}@microsoft.com,
Software Testing & Analysis (F22ST3): Static Analysis Techniques 2. Andrew Ireland
Software Testing & Analysis (F22ST3) Static Analysis Techniques Andrew Ireland School of Mathematical and Computer Science Heriot-Watt University Edinburgh Software Testing & Analysis (F22ST3): Static
Detecting Critical Defects on the Developer s Desktop
Detecting Critical Defects on the Developer s Desktop Seth Hallem CEO Coverity, Inc. Copyright Coverity, Inc. 2006. All Rights Reserved. This publication, in whole or in part, may not be reproduced, stored
Regression Verification: Status Report
Regression Verification: Status Report Presentation by Dennis Felsing within the Projektgruppe Formale Methoden der Softwareentwicklung 2013-12-11 1/22 Introduction How to prevent regressions in software
Introduction to Static Analysis for Assurance
Introduction to Static Analysis for Assurance John Rushby Computer Science Laboratory SRI International Menlo Park CA USA John Rushby Static Analysis for Assurance: 1 Overview What is static analysis?
AUTOMATED TEST GENERATION FOR SOFTWARE COMPONENTS
TKK Reports in Information and Computer Science Espoo 2009 TKK-ICS-R26 AUTOMATED TEST GENERATION FOR SOFTWARE COMPONENTS Kari Kähkönen ABTEKNILLINEN KORKEAKOULU TEKNISKA HÖGSKOLAN HELSINKI UNIVERSITY OF
Dsc+Mock: A Test Case + Mock Class Generator in Support of Coding Against Interfaces
Dsc+Mock: A Test Case + Mock Class Generator in Support of Coding Against Interfaces Mainul Islam, Christoph Csallner Computer Science and Engineering Department University of Texas at Arlington Arlington,
Taint-based Directed Whitebox Fuzzing
Taint-based Directed Whitebox Fuzzing Vijay Ganesh and Tim Leek and Martin Rinard MIT Computer Science and Artificial Intelligence Lab vganesh, tleek, [email protected] Abstract We present a new automated
Bug hunting. Vulnerability finding methods in Windows 32 environments compared. FX of Phenoelit
Bug hunting Vulnerability finding methods in Windows 32 environments compared FX of Phenoelit The goal: 0day What we are looking for: Handles network side input Runs on a remote system Is complex enough
Experimental Comparison of Concolic and Random Testing for Java Card Applets
Experimental Comparison of Concolic and Random Testing for Java Card Applets Kari Kähkönen, Roland Kindermann, Keijo Heljanko, and Ilkka Niemelä Aalto University, Department of Information and Computer
Software testing. Objectives
Software testing cmsc435-1 Objectives To discuss the distinctions between validation testing and defect testing To describe the principles of system and component testing To describe strategies for generating
GameTime: A Toolkit for Timing Analysis of Software
GameTime: A Toolkit for Timing Analysis of Software Sanjit A. Seshia and Jonathan Kotker EECS Department, UC Berkeley {sseshia,jamhoot}@eecs.berkeley.edu Abstract. Timing analysis is a key step in the
Gold Standard Method for Benchmarking C Source Code Static Analysis Tools
Gold Standard Method for Benchmarking C Source Code Static Analysis Tools Cyber Security Division 2012 Principal Investigators Meeting October 11, 2012 Henny Sipma Sr. Computer Scientist Kestrel Technology,
Combining Static Analysis and Test Generation for C Program Debugging
Combining Static Analysis and Test Generation for C Program Debugging Omar Chebaro 1,2, Nikolai Kosmatov 1, Alain Giorgetti 2,3, and Jacques Julliand 2 1 CEA, LIST, Software Safety Laboratory, PC 94, 91191
The Security Development Lifecycle
The Security Development Lifecycle Steven B. Lipner Director of Security Engineering Strategy Security Business and Technology Unit Microsoft Corporation Context and History 1960s penetrate and patch 1970s
The Security Development Lifecycle. Steven B. Lipner, CISSP [email protected] Senior Director Security Engineering Strategy Microsoft Corp.
The Security Development Lifecycle Steven B. Lipner, CISSP [email protected] Senior Director Security Engineering Strategy Microsoft Corp. 2 Overview Introduction A look back Trustworthy Computing
Minimizing code defects to improve software quality and lower development costs.
Development solutions White paper October 2008 Minimizing code defects to improve software quality and lower development costs. IBM Rational Software Analyzer and IBM Rational PurifyPlus software Kari
Fuzzing in Microsoft and FuzzGuru framework
Fuzzing in Microsoft and FuzzGuru framework OWASP-IL May-2007 John Neystadt [email protected] Lead Program Manager Forefront Edge, Microsoft Agenda Overview Introduction to Fuzzing FuzzGuru Architecture
Secure Software Programming and Vulnerability Analysis
Secure Software Programming and Vulnerability Analysis Christopher Kruegel [email protected] http://www.auto.tuwien.ac.at/~chris Testing and Source Code Auditing Secure Software Programming 2 Overview
Levels of Software Testing. Functional Testing
Levels of Software Testing There are different levels during the process of Testing. In this chapter a brief description is provided about these levels. Levels of testing include the different methodologies
find model parameters, to validate models, and to develop inputs for models. c 1994 Raj Jain 7.1
Monitors Monitor: A tool used to observe the activities on a system. Usage: A system programmer may use a monitor to improve software performance. Find frequently used segments of the software. A systems
Compositional Dynamic Test Generation
Compositional Dynamic Test Generation (Extended Abstract) Patrice Godefroid Microsoft Research [email protected] Abstract Dynamic test generation is a form of dynamic program analysis that attempts to compute
D. Best Practices D.1. Assurance The 5 th A
Best Practices I&C School Prof. P. Janson September 2014 D. Best Practices D.1. Assurance The 5 th A 1 of 20 IT systems are insecure for two main reasons: People are fallible and systems are complex and
Ivan Medvedev Principal Security Development Lead Microsoft Corporation
Ivan Medvedev Principal Security Development Lead Microsoft Corporation Session Objectives and Takeaways Session Objective(s): Give an overview of the Security Development Lifecycle Discuss the externally
Managing Latency in IPS Networks
Application Note Revision B McAfee Network Security Platform Managing Latency in IPS Networks Managing Latency in IPS Networks McAfee Network Security Platform provides you with a set of pre-defined recommended
Dynamic Test Generation To Find Integer Bugs in x86 Binary Linux Programs
Dynamic Test Generation To Find Integer Bugs in x86 Binary Linux Programs David Molnar UC Berkeley Xue Cong Li UC Berkeley David A. Wagner UC Berkeley Abstract Recently, integer bugs, including integer
Assertion Synthesis Enabling Assertion-Based Verification For Simulation, Formal and Emulation Flows
Assertion Synthesis Enabling Assertion-Based Verification For Simulation, Formal and Emulation Flows Manual Assertion Creation is ABV Bottleneck Assertion-Based Verification adopted by leading design companies
The Hacker Strategy. Dave Aitel [email protected]. Security Research
1 The Hacker Strategy Dave Aitel [email protected] Security Research Who am I? CTO, Immunity Inc. History: NSA->@stake -> Immunity Responsible for new product development Vulnerability Sharing Club
Oracle Solaris Studio Code Analyzer
Oracle Solaris Studio Code Analyzer The Oracle Solaris Studio Code Analyzer ensures application reliability and security by detecting application vulnerabilities, including memory leaks and memory access
Chapter 8 Software Testing
Chapter 8 Software Testing Summary 1 Topics covered Development testing Test-driven development Release testing User testing 2 Program testing Testing is intended to show that a program does what it is
Base One's Rich Client Architecture
Base One's Rich Client Architecture Base One provides a unique approach for developing Internet-enabled applications, combining both efficiency and ease of programming through its "Rich Client" architecture.
SAN Conceptual and Design Basics
TECHNICAL NOTE VMware Infrastructure 3 SAN Conceptual and Design Basics VMware ESX Server can be used in conjunction with a SAN (storage area network), a specialized high speed network that connects computer
Software Vulnerability Detection using Backward Trace Analysis and Symbolic Execution
Software Vulnerability Detection using Backward Trace Analysis and Symbolic Execution Hongzhe Li, Taebeom Kim, Munkhbayar Bat-Erdene and Heejo Lee Div. of Computer and Communication Engineering Korea University
Interpreters and virtual machines. Interpreters. Interpreters. Why interpreters? Tree-based interpreters. Text-based interpreters
Interpreters and virtual machines Michel Schinz 2007 03 23 Interpreters Interpreters Why interpreters? An interpreter is a program that executes another program, represented as some kind of data-structure.
The Course. http://www.cse.unsw.edu.au/~cs3153/
The Course http://www.cse.unsw.edu.au/~cs3153/ Lecturers Dr Peter Höfner NICTA L5 building Prof Rob van Glabbeek NICTA L5 building Dr Ralf Huuck NICTA ATP building 2 Plan/Schedule (1) Where and When Tuesday,
CIS570 Modern Programming Language Implementation. Office hours: TDB 605 Levine [email protected]. [email protected].
CIS570 Modern Programming Language Implementation Instructor: Admin. Assistant: URL: E Christopher Lewis Office hours: TDB 605 Levine [email protected] Cheryl Hickey [email protected] 502
CS 2112 Spring 2014. 0 Instructions. Assignment 3 Data Structures and Web Filtering. 0.1 Grading. 0.2 Partners. 0.3 Restrictions
CS 2112 Spring 2014 Assignment 3 Data Structures and Web Filtering Due: March 4, 2014 11:59 PM Implementing spam blacklists and web filters requires matching candidate domain names and URLs very rapidly
A Practical Method to Diagnose Memory Leaks in Java Application Alan Yu
A Practical Method to Diagnose Memory Leaks in Java Application Alan Yu 1. Introduction The Java virtual machine s heap stores all objects created by a running Java application. Objects are created by
Fast Arithmetic Coding (FastAC) Implementations
Fast Arithmetic Coding (FastAC) Implementations Amir Said 1 Introduction This document describes our fast implementations of arithmetic coding, which achieve optimal compression and higher throughput by
Formal Verification by Model Checking
Formal Verification by Model Checking Natasha Sharygina Carnegie Mellon University Guest Lectures at the Analysis of Software Artifacts Class, Spring 2005 1 Outline Lecture 1: Overview of Model Checking
The Advantages of Block-Based Protocol Analysis for Security Testing
The Advantages of Block-Based Protocol Analysis for Security Testing Dave Aitel Immunity,Inc. 111 E. 7 th St. Suite 64, NY NY 10009, USA [email protected] February, 4 2002 Abstract. This paper describes
RevoScaleR Speed and Scalability
EXECUTIVE WHITE PAPER RevoScaleR Speed and Scalability By Lee Edlefsen Ph.D., Chief Scientist, Revolution Analytics Abstract RevoScaleR, the Big Data predictive analytics library included with Revolution
Joeri de Ruiter Sicco Verwer
Automated reverse engineering of security protocols Learning to fuzz, fuzzing to learn Fides Aarts Erik Poll Joeri de Ruiter Sicco Verwer Radboud University Nijmegen Fuzzing 1. Plain fuzzing, with long
Symbolic Execution for Software Testing in Practice Preliminary Assessment
Symbolic Execution for Software Testing in Practice Preliminary Assessment Cristian Cadar Imperial College London [email protected] Koushik Sen U.C. Berkeley [email protected] Patrice Godefroid
How To Protect A Virtual Desktop From Attack
Endpoint Security: Become Aware of Virtual Desktop Infrastructures! An Ogren Group Special Report May 2011 Executive Summary Virtual desktops infrastructures, VDI, present IT with the unique opportunity
Abstract Interpretation-based Static Analysis Tools:
Abstract Interpretation-based Static Analysis Tools: Proving the Absence of Runtime Errors and Safe Upper Bounds on the Worst-Case Execution Time and Safe Upper Bounds on the Stack Usage Christian Ferdinand
Co-Presented by Mr. Bill Rinko-Gay and Dr. Constantin Stanca 9/28/2011
QAI /QAAM 2011 Conference Proven Practices For Managing and Testing IT Projects Co-Presented by Mr. Bill Rinko-Gay and Dr. Constantin Stanca 9/28/2011 Format This presentation is a journey When Bill and
Guideline for stresstest Page 1 of 6. Stress test
Guideline for stresstest Page 1 of 6 Stress test Objective: Show unacceptable problems with high parallel load. Crash, wrong processing, slow processing. Test Procedure: Run test cases with maximum number
How To Handle Big Data With A Data Scientist
III Big Data Technologies Today, new technologies make it possible to realize value from Big Data. Big data technologies can replace highly customized, expensive legacy systems with a standard solution
UnitCheck: Unit Testing and Model Checking Combined
UnitCheck: Unit Testing and Model Checking Combined Michal Kebrt and Ondřej Šerý Charles University in Prague Malostranské náměstí 25 118 00 Prague 1 Czech Republic [email protected], [email protected]
The programming language C. sws1 1
The programming language C sws1 1 The programming language C invented by Dennis Ritchie in early 1970s who used it to write the first Hello World program C was used to write UNIX Standardised as K&C (Kernighan
Cloud9 Parallel Symbolic Execution for Automated Real-World Software Testing
Cloud9 Parallel Symbolic Execution for Automated Real-World Software Testing Stefan Bucur, Vlad Ureche, Cristian Zamfir, George Candea School of Computer and Communication Sciences Automated Software Testing
Static Analysis. Find the Bug! 15-654: Analysis of Software Artifacts. Jonathan Aldrich. disable interrupts. ERROR: returning with interrupts disabled
Static Analysis 15-654: Analysis of Software Artifacts Jonathan Aldrich 1 Find the Bug! Source: Engler et al., Checking System Rules Using System-Specific, Programmer-Written Compiler Extensions, OSDI
Peach Fuzzer Platform
Fuzzing is a software testing technique that introduces invalid, malformed, or random data to parts of a computer system, such as files, network packets, environment variables, or memory. How the tested
Module 10. Coding and Testing. Version 2 CSE IIT, Kharagpur
Module 10 Coding and Testing Lesson 26 Debugging, Integration and System Testing Specific Instructional Objectives At the end of this lesson the student would be able to: Explain why debugging is needed.
How To Trace
CS510 Software Engineering Dynamic Program Analysis Asst. Prof. Mathias Payer Department of Computer Science Purdue University TA: Scott A. Carr Slides inspired by Xiangyu Zhang http://nebelwelt.net/teaching/15-cs510-se
Dynamic Resource management with VM layer and Resource prediction algorithms in Cloud Architecture
Dynamic Resource management with VM layer and Resource prediction algorithms in Cloud Architecture 1 Shaik Fayaz, 2 Dr.V.N.Srinivasu, 3 Tata Venkateswarlu #1 M.Tech (CSE) from P.N.C & Vijai Institute of
Visualizing Information Flow through C Programs
Visualizing Information Flow through C Programs Joe Hurd, Aaron Tomb and David Burke Galois, Inc. {joe,atomb,davidb}@galois.com Systems Software Verification Workshop 7 October 2010 Joe Hurd, Aaron Tomb
Software Engineering Introduction & Background. Complaints. General Problems. Department of Computer Science Kent State University
Software Engineering Introduction & Background Department of Computer Science Kent State University Complaints Software production is often done by amateurs Software development is done by tinkering or
The software model checker BLAST
Int J Softw Tools Technol Transfer (2007) 9:505 525 DOI 10.1007/s10009-007-0044-z SPECIAL SECTION FASE 04/05 The software model checker BLAST Applications to software engineering Dirk Beyer Thomas A. Henzinger
Software safety - DEF-STAN 00-55
Software safety - DEF-STAN 00-55 Where safety is dependent on the safety related software (SRS) fully meeting its requirements, demonstrating safety is equivalent to demonstrating correctness with respect
Heuristics for Dynamically Adapting Constraint Propagation in Constraint Programming
Heuristics for Dynamically Adapting Constraint Propagation in Constraint Programming Kostas Stergiou AI Lab University of the Aegean Greece CPAIOR 09 Workshop on Bound reduction techniques for CP and MINLP
Automated Program Behavior Analysis
Automated Program Behavior Analysis Stacy Prowell [email protected] March 2005 SQRL / SEI Motivation: Semantics Development: Most engineering designs are subjected to extensive analysis; software is
what operations can it perform? how does it perform them? on what kind of data? where are instructions and data stored?
Inside the CPU how does the CPU work? what operations can it perform? how does it perform them? on what kind of data? where are instructions and data stored? some short, boring programs to illustrate the
ByteMobile Adaptive Traffic Management Product Family
ByteMobile Adaptive Traffic Management Product Family Building Adaptive Traffic Management Solutions ByteMobile Adaptive Traffic Management Solutions allow mobile operators to actively and dynamically
Basics of VTune Performance Analyzer. Intel Software College. Objectives. VTune Performance Analyzer. Agenda
Objectives At the completion of this module, you will be able to: Understand the intended purpose and usage models supported by the VTune Performance Analyzer. Identify hotspots by drilling down through
Binary search tree with SIMD bandwidth optimization using SSE
Binary search tree with SIMD bandwidth optimization using SSE Bowen Zhang, Xinwei Li 1.ABSTRACT In-memory tree structured index search is a fundamental database operation. Modern processors provide tremendous
How To Test A Web Server
Performance and Load Testing Part 1 Performance & Load Testing Basics Performance & Load Testing Basics Introduction to Performance Testing Difference between Performance, Load and Stress Testing Why Performance
Delivering Quality in Software Performance and Scalability Testing
Delivering Quality in Software Performance and Scalability Testing Abstract Khun Ban, Robert Scott, Kingsum Chow, and Huijun Yan Software and Services Group, Intel Corporation {khun.ban, robert.l.scott,
Part V Applications. What is cloud computing? SaaS has been around for awhile. Cloud Computing: General concepts
Part V Applications Cloud Computing: General concepts Copyright K.Goseva 2010 CS 736 Software Performance Engineering Slide 1 What is cloud computing? SaaS: Software as a Service Cloud: Datacenters hardware
Wiggins/Redstone: An On-line Program Specializer
Wiggins/Redstone: An On-line Program Specializer Dean Deaver Rick Gorton Norm Rubin {dean.deaver,rick.gorton,norm.rubin}@compaq.com Hot Chips 11 Wiggins/Redstone 1 W/R is a Software System That: u Makes
Secure in 2010? Broken in 2011!
Secure in 2010? Broken in 2011! Matias Madou Principal Security Researcher Abstract In 2010, a security research firm stumbled on a couple of vulnerabilities in Apache OFBiz, a widely used open source
Scalable Data Analysis in R. Lee E. Edlefsen Chief Scientist UserR! 2011
Scalable Data Analysis in R Lee E. Edlefsen Chief Scientist UserR! 2011 1 Introduction Our ability to collect and store data has rapidly been outpacing our ability to analyze it We need scalable data analysis
Loop-Extended Symbolic Execution on Binary Programs
Loop-Extended Symbolic Execution on Binary Programs Prateek Saxena Pongsin Poosankam Stephen McCamant Dawn Song University of California, Berkeley and Carnegie Mellon University Berkeley, California, USA
How To Use Windows Small Business Server 2011 Essentials
Everything Your Business Needs in a Server, Nothing it doesn t. Ideal as a first server for small businesses with up to 25 users, Windows Small Business Server 2011 Essentials provides a cost-effective
Two Flavors in Automated Software Repair: Rigid Repair and Plastic Repair
Two Flavors in Automated Software Repair: Rigid Repair and Plastic Repair Martin Monperrus, Benoit Baudry Dagstuhl Preprint, Seminar #13061, 2013. Link to the latest version Abstract In this paper, we
