Serena Business Manager Performance Test Results



Similar documents
SERENA SOFTWARE Providing Custom Documentation for Serena Business Manager

SERENA SOFTWARE Serena Service Manager Security

SERENA SOFTWARE Authors: Bill Weingarz, Pete Dohner, Kartik Raghavan, Amitav Chakravartty

SERENA SOFTWARE Authors: Bill Weingarz, Pete Dohner, Kartik Raghavan, Amitav Chakravartty

SOLUTION BRIEF: SLCM R12.8 PERFORMANCE TEST RESULTS JANUARY, Submit and Approval Phase Results

Business Application Services Testing

Fundamentals of LoadRunner 9.0 (2 Days)

SOLUTION BRIEF: SLCM R12.7 PERFORMANCE TEST RESULTS JANUARY, Load Test Results for Submit and Approval Phases of Request Life Cycle

Performance Testing Process A Whitepaper

Contents. Serena Business Manager 2009 R4.02 Readme. Turn On Tabs

PC120 ALM Performance Center 11.5 Essentials

SysPatrol - Server Security Monitor

1. Welcome to QEngine About This Guide About QEngine Online Resources Installing/Starting QEngine... 5

HPE PC120 ALM Performance Center 12.0 Essentials

Informatica Master Data Management Multi Domain Hub API: Performance and Scalability Diagnostics Checklist

efolder BDR for Veeam Cloud Connection Guide

An Oracle White Paper July Oracle Primavera Contract Management, Business Intelligence Publisher Edition-Sizing Guide

Monitoring HP OO 10. Overview. Available Tools. HP OO Community Guides

Lepide Active Directory Self Service. Configuration Guide. Follow the simple steps given in this document to start working with

Aradial Installation Guide

Diagnostic Manager. User Guide. Publication Date: September 04, 2015

FileMaker Server 7. Administrator s Guide. For Windows and Mac OS

Chapter 8 Monitoring and Logging

Oracle Hyperion Financial Management Virtualization Whitepaper

vcenter Operations Management Pack for SAP HANA Installation and Configuration Guide

An Oracle White Paper March Load Testing Best Practices for Oracle E- Business Suite using Oracle Application Testing Suite

How To Test On The Dsms Application

HP LoadRunner. Software Version: LoadRunner Tutorial

Installation and Configuration Guide for Windows and Linux

SOFTWARE TESTING TRAINING COURSES CONTENTS

NTP Software File Auditor for Windows Edition

ADAM 5.5. System Requirements

Server Installation Manual 4.4.1

NTP Software QFS for NAS, NetApp Edition Installation Guide

Integration Service Database. Installation Guide - Oracle. On-Premises

JBoss Seam Performance and Scalability on Dell PowerEdge 1855 Blade Servers

Testhouse Training Portfolio

SAM Server Utility User s Guide

Load Testing Hyperion Applications Using Oracle Load Testing 9.1

System Administration of Windchill 10.2

StreamServe Persuasion SP5 StreamStudio

Copyright 2012 Trend Micro Incorporated. All rights reserved.

SQL Server Instance-Level Benchmarks with DVDStore

Table of Contents INTRODUCTION Prerequisites... 3 Audience... 3 Report Metrics... 3

Veeam Backup Enterprise Manager. Version 7.0

serena.com Best Practice for Parallel Development using Serena Dimensions

Exchange Administrators will be able to use a more secure authentication mechanism compared with username and password

About This Guide Signature Manager Outlook Edition Overview... 5

MySQL Enterprise Monitor

Troubleshooting problems with the PDMWorks Enterprise database server

NETWRIX EVENT LOG MANAGER

Enterprise Manager. Version 6.2. Administrator s Guide

Installation and Configuration Guide for Windows and Linux

Sisense. Product Highlights.

Departmental (Service) Account Set Up

Getting Started with SandStorm NoSQL Benchmark

Benchmarking Guide. Performance. BlackBerry Enterprise Server for Microsoft Exchange. Version: 5.0 Service Pack: 4

MONITORING PERFORMANCE IN WINDOWS 7

Upgrading from Call Center Reporting to Reporting for Contact Center. BCM Contact Center

A closer look at HP LoadRunner software

User s Manual. Management Software for ATS

Running a Workflow on a PowerCenter Grid

Identifying Performance Bottleneck using JRockit. - Shivaram Thirunavukkarasu Performance Engineer Wipro Technologies

Usage Analysis Tools in SharePoint Products and Technologies

HP ALM. Software Version: Tutorial

Sage Intelligence Financial Reporting for Sage ERP X3 Version 6.5 Installation Guide

TNT SOFTWARE White Paper Series

Estimate Performance and Capacity Requirements for Workflow in SharePoint Server 2010

An Oracle White Paper September Advanced Java Diagnostics and Monitoring Without Performance Overhead

CoCreate Manager Server Installation Guide. CoCreate Manager Server Installation Guide 1

Performance Monitor. Intellicus Web-based Reporting Suite Version 4.5. Enterprise Professional Smart Developer Smart Viewer

An Introduction to LoadRunner A Powerful Performance Testing Tool by HP. An Introduction to LoadRunner. A Powerful Performance Testing Tool by HP

PERFORMANCE TESTING. New Batches Info. We are ready to serve Latest Testing Trends, Are you ready to learn.?? START DATE : TIMINGS : DURATION :

Symantec Endpoint Protection Shared Insight Cache User Guide


Installation Guide. MashZone. Version 9.6

LR120 LoadRunner 12.0 Essentials

NetIQ Access Manager 4.1

SMock A Test Platform for the Evaluation of Monitoring Tools

FileMaker 12. ODBC and JDBC Guide

HCIbench: Virtual SAN Automated Performance Testing Tool User Guide

Enterprise Manager Performance Tips

Orchestrated. Release Management. Gain insight and control, eliminate ineffective handoffs, and automate application deployments

Performance Testing of a Cloud Service

Intellicus Enterprise Reporting and BI Platform

Adept 2014 System Requirements

Profiling and Testing with Test and Performance Tools Platform (TPTP)

Hardware Recommendations

Embarcadero DB Change Manager 6.0 and DB Change Manager XE2

Performance And Scalability In Oracle9i And SQL Server 2000

LR120 Load Runner 12.0 Essentials Instructor-Led Training Version 12.0

18.2 user guide No Magic, Inc. 2015

AT&T Toggle. 4/23/2014 Page i

WEBAPP PATTERN FOR APACHE TOMCAT - USER GUIDE

Justin Bruns, Performance Test Lead, Fireman's Fund Insurance Robert May, Enterprise Software Tools Administrator Fireman's Fund Insurance

ARIS Server Installation and Administration Guide ARIS. Version Service Release 1

Oracle Primavera P6 Enterprise Project Portfolio Management Performance and Sizing Guide. An Oracle White Paper October 2010

Tuning WebSphere Application Server ND 7.0. Royal Cyber Inc.

Drobo How-To Guide. What You Will Need. Configure Replication for DR Using Double-Take Availability and Drobo iscsi SAN

Transcription:

SERENA SOFTWARE Serena Business Manager Performance Test Results Author: RT Tangri 2015-04-09

Table of Contents Who Should Read This Paper?... 3 Test Methodology... 3 Runtime Test Architecture... 4 Load Test Scenario... 4 About the Data Set... 7 Performance Test Results... 8 Additional Information... 21 No Title... 0 2 Serena Software

Who Should Read This Paper? This paper is intended for system administrators or others interested in performance test results for Serena Business Manager 10.1.5. This paper discusses the performance tests results of 200 virtual users in the onpremise version of Serena Work Center using Windows 2008 R2 and Oracle 11gR2. Test Methodology Testing was conducted in a private enterprise performance testing lab using HP's LoadRunner 12.00. Tests measured application response time, throughput, and system resource usage under a load of 200 virtual users against a large enterprise data set with a think time pause of 10 seconds between transactions. This test involved making Load Runner generate load by creating virtual users (one unique virtual user is added every 5 seconds until the desired load level is reached) to model user activity. Each virtual user then performs scripted tasks and sends crafted HTTP requests. LoadRunner then uses a threading execution model to create multiple instances of unique virtual users to create load and concurrency on the SBM application. During this test, the 200 virtual users iterate 600 times through each use case (transaction) in the performance suite. The scripts and the datasets used in this testing can be provided for reference upon request. Serena Business Manager Performance Test Results 3

Runtime Test Architecture Runtime Test Architecture The performance test was performed in a distributed SBM installation using 5 separate 64-bit servers. The environment included a dedicated Serena License Manager installation with 80,000 seat licenses. Server 1 SBM Application Engine Web Server (Win2008R2-IIS 7 or Win2012R2-IIS 8.5) with 8GB memory Server 2 SBM JBoss1: Single Sign-On (SSO), SBM Mail Services, SBM Orchestration Engine with 8GB total and 4GB memory allocated to JBoss Server 3 SBM JBoss2: SBM Common Services (Smart Search, Resource Mgmt, Relationship Service, SSF Services) with 8GB total and 4GB memory allocated to JBoss Server 4 SBM Logging Services (Mongod64) with 8GB of total memory Server 5 Oracle 11gR2 database with 16GB of total memory Note: In 10.1.5, SBM Configurator enables you to install SBM Logging Services (MongoDB) on a separate server. As part of the performance tests, SBM Logging Services was installed on a separate server to measure its full out-of-the-box impact at 200vu load. This configuration is typically recommended only if you plan to set the logging level to TRACE for debugging purposes. Load Test Scenario Load tests simulated scaling the number of common tasks that are performed by 200 virtual users. Using a standard defect tracking application that enables a team to track product defects, the tests replicated the life cycle of defects from submittal, approval, and assignment. Notes and attachments are added to each item as it moves through the process. The Notification Server sent e-mails via SMTP to a test SMTP Mail Server repository. Shared views of activity, backlog, calendar feeds, and shared dashboard views of multiview reports were run against the system. 4 Serena Software

Serena Work Center (SBM Application Engine) Use Cases Tests were performed by a set number of 200 unique virtual users over a 20 minute baseline. Think-time between each use case (LoadRunner Transaction ) was 10 seconds (with exceptions for transactions from Serena Work Center: 60 second think-time before indexer transactions; 90 second think-time before notification transactions). The performance workflow consist of the following actions: LoadRunner virtual users run in multiple threads of execution 200 unique virtual users log in via SSO to Serena Work Center in vuser_init Repeat iterate (defect tracking application) use cases 200 unique virtual users log off via SSO from Serena Work Center in vuser_end SERENA WORK CENTER (SBM APPLICATION ENGINE) USE CASES Serena Work Center use cases run during virtual user initialization against large enterprise data set: Web SSO Login VUser Init Notification Remove All Settings Save Preferred Project Settings Save Items Per Page 40 And Notification Poll Interval 90 Second SWC Home Add Report Widget (Of Multi View Report #1) SWC Home Add Activity Widget (My Activity) Serena Work Center use cases run during virtual user exit against large enterprise data set: SWC Home Delete Report Widget (Of Multi View Report #1) SWC Home Delete Activity Widget (My Activity) VUser End Notification Remove All Web SSO Logout Serena Work Center use cases iterated by each virtual user against large enterprise data set: Submit-Form Submit-Form OK (Submit Item) Transition Owner In Defect Tracking Activities Tab Defect Tracking My Activity After Owner IN Defect Tracking Shared Activity Feed View After Owner IN Defect Tracking Calendars Tab Defect Tracking My Calendar Defect Tracking Shared Calendar Feed View After Owner IN Transition Owner Out Transition Owner Out Ok Transition Secondary Owner In My Activity After Secondary Owner In Shared Activity View All Items I Own Primary and Secondary Transition Secondary Owner Out Transition Secondary Owner Out Ok Transition-SyncTest (makes Synchronous AE-WebServices call) AddNote AddNote-Ok (Text with HTML5 Tags Includes 27KB JPG + 2KB Note) AddAttachment AddAttachment-OK (11KB TXT File Attachment) Serena Business Manager Performance Test Results 5

Load Test Scenario Email_Link Email_Link SearchUser Email_Link-OK (Message with HTML5 Tags) Transition Assign To CCB Transition Assign To CCB OK Transition Assign To Area Owner Transition Assign To Area Owner OK Social View Of BUG BUGID Lucene Text Search BUGStar Lucene Text Search BUGIDStar Lucene Text Search Attachment Lucene Text Search TitleDescription Lucene Text Search TitleDescriptionStar Lucene Text Search Submitter Lucene Text Search User Profile Card Add then Delete From Favorites PinUp Defect Tracking (Application) Select Defect Tracking PinUp Defect Tracking Manage Views Defect Tracking My Dashboard Defect Tracking Shared Dashboard View (Of Multi View Report #1) Defect Tracking Backlogs Tab Defect Tracking Shared Backlog Feed View Report Center CTOSDS Elapsed Duration CTOSDS in State Duration (Excluded from Oracle PERF Run Due To High Response) CTOSDS Average To State Duration (Excluded from Oracle PERF Run Due To High Response) CTOSDS Open And Completed Trend (Excluded from Oracle PERF Run Due To High Response) CTOSDS Entering A State Trend CTOSDS State Activity Trend (Excluded from Oracle PERF Run Due To High Response) Multi-View Report 1 Static Listing For Performance EG Static Listing For Multi-View Performance All Issues By Project and State All Issues by Issue Type Listing Report All Open Issues My Open Issues All Issues by Project and Issue Type All Issues by State All Issues by Owner Notification UI View All Notification UI MarkRead and Remove N (For Each Notification Object) Each Iteration Notification Remove All 6 Serena Software

Serena Work Center (SBM Application Engine) Use Cases About the Data Set The large enterprise data set that was used for testing is summarized below. SBM Entity Count Workflows (define the process that items follow) 31 Projects (store process items, such as issues and incidents) 5,302 Users 21,100 SBM groups 138 Folders (store personal and group favorites) 147,570 Issues (process items that follow a workflow) 1,130,000 Contacts (similar to address book entries) 20,053 Note: In addition, the database included the following configuration settings: processes = 300 (Newer versions of Application Engine require more sessions per user to power Work Center) READ_COMMITTED_SNAPSHOT = OFF Serena Business Manager Performance Test Results 7

Performance Test Results Performance Test Results The following graphs summarize the results from the performance tests. AVERAGE TRANSACTION RESPONSE TIME The response time for individual transactions are summarized below. Measurement Minimum Average Maximum SWC Home Delete Report Widget 0.035.04 0.045 VUser End Notification Remove All 0.035 0.042 0.052 SWC Home Delete Activity Widget 0.036 0.043 0.064 BUGID Star Lucene Text Search 0.038 0.044 0.054 Add Attachment 0.045 0.053 0.069 User Profile Card 0.05 0.06 0.095 Defect Tracking Backlogs Tab 0.068 0.074 0.09 Defect Tracking My Calendar 0.072 0.082 0.111 Title Description Lucene Text Search 0.067 0.093 0.161 Email Link Search User 0.085 0.095 0.113 8 Serena Software

Average Transaction Response Measurement Minimum Average Maximum Add Note 0.089 0.103 0.141 Defect Tracking Calendars Tab 0.092 0.108 0.173 Listing Report 0.093 0.11 0.137 My Open Issues 0.106 0.12 0.154 Email Link 0.107 0.124 0.157 Defect Tracking Activities Tab 0.1 0.125 0.152 CTOSDS in State Duration 0.117 0.133 0.166 Title Description Star Lucene Text Search 0.099 0.134 0.193 Add then Delete From Favorites 0.13 0.146 0.169 Notification UI View All 0.157 0.179 0.225 Notification UI Mark Read and Remove 1 0.165 0.193 0.222 Notification UI Mark Read and Remove 10 0.171 0.194 0.235 Notification UI Mark Read and Remove 9 0.168 0.194 0.246 Notification UI Mark Read and Remove 2 0.172 0.196 0.224 Notification UI Mark Read and Remove 3 0.172 0.196 0.228 Notification UI Mark Read and Remove 4 0.178 0.196 0.232 Notification UI Mark Read and Remove 8 0.17 0.196 0.233 Notification UI Mark Read and Remove 7 0.168 0.197 0.237 Notification UI Mark Read and Remove 5 0.176 0.199 0.246 Notification UI Mark Read and Remove 6 0.173 0.201 0.244 BUGID Lucene Text Search 0.042 0.204 0.57 Settings Save Items Per Page 40 And Notification Poll Interval 90 Second 0.182 0.211 0.272 VUser Init Notification Remove All 0.043 0.23 0.453 All Open Issues 0.226 0.253 0.299 Serena Business Manager Performance Test Results 9

Performance Test Results Measurement Minimum Average Maximum Email Link OK 0.241 0.262 0.283 Defect Tracking Manage Views 0.242 0.266 0.298 Static Listing For Multi-View Performance 0.256 0.28 0.31 Social View Of BUG 0.26 0.3 0.389 Select Defect Tracking Pin Up 0.29 0.324 0.391 Defect Tracking Shared Calendar Feed View After Owner IN 0.265 0.325 0.385 Static Listing For Performance EG 0.299 0.335 0.388 Submit Form 0.305 0.336 0.413 Defect Tracking My Dashboard 0.328 0.359 0.398 Transition Secondary Owner OUT 0.344 0.381 0.412 Transition Owner OUT 0.343 0.384 0.447 Each Iteration Notification Remove All 0.305 0.385 0.535 Transition Assign To CCB 0.345 0.387 0.431 All Issues by Project and Issue Type 0.345 0.406 0.478 All Issues by Issue Type 0.367 0.417 0.499 Transition Assign To Area Owner 0.394 0.419 0.488 All Issues by State 0.361 0.427 0.507 Transition Owner OUT OK 0.386 0.433 0.497 Transition Secondary Owner OUT OK 0.404 0.433 0.48 All Issues by Owner 0.364 0.435 0.747 All Issues By Project and State 0.392 0.457 0.508 Defect Tracking My Activity After Owner IN 0.434 0.474 0.519 Add Attachment OK 0.444 0.477 0.536 My Activity After Secondary Owner IN 0.411 0.478 0.525 10 Serena Software

Average Transaction Response Measurement Minimum Average Maximum Pin Up Defect Tracking 0.461 0.502 0.552 Transition Owner IN 0.494 0.53 0.581 Defect Tracking Shared Activity Feed View After Owner IN 0.467 0.534 0.618 Transition Assign To CCB OK 0.504 0.553 0.609 Transition Secondary Owner IN 0.518 0.554 0.605 Web SSO Logout 0.536 0.574 0.626 Submit Form OK 0.581 0.625 0.744 Add Note OK 0.595 0.628 0.669 Report Center 0.46 0.631 0.788 Transition Assign To Area Owner OK 0.6 0.637 0.724 Shared Activity View All Items I Own Primary and Secondary 0.609 0.711 0.858 CTOSDS Entering A State Trend 0.71 0.807 0.979 Transition Sync Test 0.782 0.842 0.945 CTOSDS Elapsed Duration 0.848 0.963 1.131 Multi-View Report 1 1.001 1.122 1.24 Defect Tracking Shared Dashboard View 0.974 1.13 1.239 Settings Save Preferred Project 1.02 1.265 1.88 Web SSO Login 1.112 1.41 2.917 Defect Tracking Shared Backlog Feed View 1.223 1.439 2.969 SWC Home Add Report Widget 1.496 1.66 1.973 Attachment Lucene Text Search 1.815 2.053 2.198 BUG Star Lucene Text Search 2.03 2.254 2.542 SWC Home Add Activity Widget 2.544 2.684 2.818 Serena Business Manager Performance Test Results 11

Performance Test Results Measurement Minimum Average Maximum Submitter Lucene Text Search 4.483 4.74 4.852 A higher average response time was recorded for searches with an Oracle database; for all other transactions, the average response time was 2.6 seconds or less. Search performance was impacted by high CPU/memory usage due to SLS calls to the Java Localization Service (see DEF263647). No significant performance impact was observed when the Rich Text Editor was enabled for Memo/ Note Fields in submit and transition forms. TRANSACTION HITS PER SECOND Hits per second for HTTP transactions that were executed concurrently in SBM are summarized below. 12 Serena Software

Transaction Bytes Per Second TRANSACTION BYTES PER SECOND Bytes per second for HTTP transactions that were executed concurrently in SBM are summarized below. NUMBER OF TRANSACTIONS EXECUTED CONCURRENTLY BY 200 VIRTUAL USER LOAD The number of HTTP transactions that were executed concurrently in SBM are summarized below. Serena Business Manager Performance Test Results 13

Performance Test Results WEB SERVER CPU USAGE Web server CPU usage is summarized below. WEB SERVER MEMORY USAGE Web server memory usage is summarized below. 14 Serena Software

Web Server Thread Count WEB SERVER THREAD COUNT Web server thread count is summarized below. WEB SERVER HANDLE COUNT Web server handle count is summarized below. Serena Business Manager Performance Test Results 15

Performance Test Results WEB SERVER VIRTUAL BYTES Web server virtual bytes are summarized below. COMMON SERVICES CPU USAGE Common Services CPU usage is summarized below. 16 Serena Software

Common Services Memory Usage COMMON SERVICES MEMORY USAGE Common Services memory usage is summarized below. NOTIFICATION SERVER CPU USAGE Notification Server CPU usage is summarized below. Serena Business Manager Performance Test Results 17

Performance Test Results NOTIFICATION SERVER MEMORY USAGE Notification Server memory usage is summarized below. DATABASE SERVER CPU USAGE Database Server CPU usage is summarized below. 18 Serena Software

Database Server Memory Usage DATABASE SERVER MEMORY USAGE Database Server memory usage is summarized below. SBM LOGGING SERVICES CPU USAGE SBM Logging Services (mongod64) CPU usage is summarized below. Note that Active Diagnostics logging level was set to INFO. Serena Business Manager Performance Test Results 19

Performance Test Results SBM LOGGING SERVICES MEMORY USAGE SBM Logging Services memory usage is summarized below. Note that Active Diagnostics logging level was set to INFO. RELATIONSHIP SERVICE PERFORMANCE Results for one Neo4J instance (not clustered), JBoss 3GB Heap + Neo4J 1GB Heap, default 10 Worker Threads are summarized below. Neo4J Java CPU% peak was 25% during the initial load (most CPU intensive period). Initial load totaled 1.2 million records (ExportToCSV + ImportToGraphDB), which took 108 minutes to finish. Replicating/synchronization of new records in TS_CHANGEACTIONS table was processed at a rate of 2,400 records per minute. For 1,137,446 records (and 14,118,424 relationships), Neo4J database disk usage was 3.9GB. SMART SEARCH INDEXER PERFORMANCE Results of the Smart Search indexer (Lucene), 4 GB Common Services Java Heap are summarized below. Initial index or re-index operations are now multi-threaded. 6 worker threads of execution quickly processed all primary tables sequentially. After the initial index or re-index finished, the indexer periodically polled the database every 30 seconds (by default) to update the index with new records from the TS_CHANGEACTIONS table, which was then searchable via Serena Work Center. Initial index of 1,102,276 items in the TTT_ISSUES table took 2 hours and 6 Minutes to complete (indexdir folder size was 230MB). Impact of initial index of 1,102,276 items in TTT_ISSUES: Common Services Java CPU increased to 18%, Common Services Java Working Set increased to 720MB. The indexer finished the update of 46,357 change items in 17 minutes and 45 seconds. Impact of update index with new records from TS_CHANGEACTIONS table: Common Services Java CPU increased to 2%, Common Services Java Working Set increased to 457MB. 20 Serena Software

Configuration Notes Additional Information Note the following performance testing information. CONFIGURATION NOTES If TRACE level logging in Active Diagnostics is enabled for debugging SBM components on a 64-bit system with 8GB RAM, the mongod64 process will compete for memory with SBM Common Services and prevent JBoss from using the 4GB Heap that is allocated to it. The default performance settings in SBM Configurator are used for load tests: Notification Server ForceSend = OFF Throttling of Orchestrations = OFF On the server that hosts SBM Common Services, the indexer settings are changed for load tests in SBM Configurator: Expected Search Usage = High Database Polling Interval = 30 seconds Oracle ODBC client connection to Oracle server configuration tips: Another variable for number of sessions used depends on what type of ODBC connection Application Engine is making as a client to the Oracle 11gR2 server. ODBC connection as Client10 (SSL) or via Oracle Advanced Security requires an increase in processes/sessions on theoracle 11gR2 server. To avoid server refused the connection errors in the Application Engine Event Viewer and JBDC connection exceptions in the JBoss server.log in the performance lab at 200vu load against Work Center, increasing the number of processes to 300 and sessions to 400 in Oracle is sufficient: sqlplus system/testauto@hpq5 > alter system set processes=300 scope=spfile; Verify change: sqlplus system/testaut0@hpq5 > show parameters processes; > show parameters sessions; SBM 10.1.5 REPORTED DEFECTS The following defects have been submitted against the GA build of SBM 10.1.5. DEF262318 SWC PERF Oracle11GR2 Need Open And Completed Trend Report PERF Improvement DEF262744 SWC PERF Oracle11GR2 Need Average To State Duration Report PERF Improvement DEF262748 SWC PERF Oracle11GR2 Need In State Duration Report PERF Improvement DEF262793 SWC PERF Oracle11GR2 Need State Activity Trend Report PERF Improvement DEF263647 200vu Lucene Search results comparatively slower due to MLS usage DEF261371 SWC PERF 200vu Default W3WP Process Handle Count Leak PERFORMANCE TEST EXCLUSIONS The following items were excluded from performance testing. System report transactions are not included in 200vu load test. Serena Business Manager Performance Test Results 21

Additional Information High-response duration and trend reports were excluded from the Oracle performance test run due to inconsistent results with the current Data Direct ODBC drivers. ABOUT SERENA Serena Software, Inc. provides Orchestrated IT solutions to the Global 2000. Serena's core purpose is to advance the business value of IT. Our 4,000 active enterprise customers, encompassing almost one million users worldwide, have made Serena the largest independent ALM vendor and the only one that orchestrates DevOps, the processes that bring together application development and operations. Headquartered in Silicon Valley, Serena serves enterprise customers from 29 offices in 14 countries. Serena is a portfolio company of HGGC. CONTACT Website: http://www.serena.com Copyright 2015 Serena Software, Inc.. All rights reserved. Serena, TeamTrack, ChangeMan, PVCS, StarTool, Collage, and Comparex are registered trademarks of Serena Software, Inc. Change Governance, Command Center, Dimensions, Mover and Composer are trademarks of Serena Software, Inc.. All other product or company names are used for identification purposes only, and may be trademarks of their respective owners. Revised 12 February 2015. 22 Serena Software