2013 Thomas Skybakmoen, Francisco Artes, Bob Walder, Ryan Liles



Similar documents
2013 Thomas Skybakmoen, Francisco Artes, Bob Walder, Ryan Liles

NEXT GENERATION FIREWALL COMPARATIVE ANALYSIS

How To Create A Firewall Security Value Map (Svm) 2013 Nss Labs, Inc.

SSL Performance Problems

DATA CENTER IPS COMPARATIVE ANALYSIS

How To Get A Fortinet Security System For Free

WEB APPLICATION FIREWALL COMPARATIVE ANALYSIS

TEST METHODOLOGY. Hypervisors For x86 Virtualization. v1.0

DATA CENTER IPS COMPARATIVE ANALYSIS

TEST METHODOLOGY. Network Firewall Data Center. v1.0

WEB APPLICATION FIREWALL PRODUCT ANALYSIS

ENTERPRISE EPP COMPARATIVE REPORT

Internet Advertising: Is Your Browser Putting You at Risk?

NEXT GENERATION FIREWALL COMPARATIVE ANALYSIS

TEST METHODOLOGY. Web Application Firewall. v6.2

Breach Found. Did It Hurt?

TEST METHODOLOGY. Data Center Firewall. v2.0

NETWORK FIREWALL TEST METHODOLOGY 3.0. To receive a licensed copy or report misuse, Please contact NSS Labs at: or advisor@nsslabs.

How To Test A Ddos Prevention Solution

Can Consumer AV Products Protect Against Critical Microsoft Vulnerabilities?

NETWORK INTRUSION PREVENTION SYSTEM PRODUCT ANALYSIS

Is the Security Industry Ready for SSL Decryption?

Evolutions in Browser Security

TEST METHODOLOGY. Distributed Denial-of-Service (DDoS) Prevention. v2.0

NEXT GENERATION FIREWALL TEST REPORT

NEXT GENERATION FIREWALL PRODUCT ANALYSIS

Internet Explorer Exploit Protection ENTERPRISE BRIEFING REPORT

An Old Dog Had Better Learn Some New Tricks

CORPORATE AV / EPP COMPARATIVE ANALYSIS

NEXT GENERATION FIREWALL PRODUCT ANALYSIS

4 Delivers over 20,000 SSL connections per second (cps), which

NEXT GENERATION INTRUSION PREVENTION SYSTEM (NGIPS) TEST REPORT

BROWSER SECURITY COMPARATIVE ANALYSIS

The CISO s Guide to the Importance of Testing Security Devices

Mobile App Containers: Product Or Feature?

Multiple Drivers For Cyber Security Insurance

Performance of Cisco IPS 4500 and 4300 Series Sensors

Symantec Endpoint Protection 11.0 Architecture, Sizing, and Performance Recommendations

Next-Generation Firewalls: CEO, Miercom

Why Is DDoS Prevention a Challenge?

ACHILLES CERTIFICATION. SIS Module SLS 1508

Comparative Performance and Resilience Test Results - UTM Appliances. Miercom tests comparing Sophos SG Series appliances against the competition

NETWORK INTRUSION PREVENTION SYSTEM

LOAD BALANCING IN THE MODERN DATA CENTER WITH BARRACUDA LOAD BALANCER FDC T740

How To Sell Security Products To A Network Security Company

Milestone Solution Partner IT Infrastructure MTP Certification Report Scality RING Software-Defined Storage

SuperGIS Server 3 Website Performance and Stress Test Report

High Performance NGFW Extended

Web Security Firewall Setup. Administrator Guide

Achieve Deeper Network Security

Dell One Identity Cloud Access Manager How to Configure for High Availability

Next-Generation Firewalls: Critical to SMB Network Security

NETWORK INTRUSION PREVENTION SYSTEM

How to Build a Massively Scalable Next-Generation Firewall

Customer Advantage Program

Dell Advanced Network Monitoring Services Service Description

Network Security Equipment The Ever Changing Curveball

Dell One Identity Cloud Access Manager How to Configure for SSO to SAP NetWeaver using SAML 2.0

Frequently Asked Questions

DATA$CENTER$FIREWALL$PRODUCT$ANALYSIS$$

Windows Hard Disk Encryption

WAN Optimization, Web Cache, Explicit Proxy, and WCCP. FortiOS Handbook v3 for FortiOS 4.0 MR3

Achieve Deeper Network Security and Application Control

Network Simulation Traffic, Paths and Impairment

Secure Upgrade Plus. Customer Advantage Program. Secure Upgrade Plus. Customer Loyalty Bundles. Address other business needs

User Manual. Page 2 of 38

IBM Proventia Network Intrusion Prevention System With Crossbeam X80 Platform

COM Port Stress Test

Unified Threat Management Throughput Performance

ZXTM (Zeus Extensible Traffic Manager) In Virtual Mode With Microsoft Virtual Server

Evaluating Wireless Broadband Gateways for Deployment by Service Provider Customers

FortiAuthenticator Agent for Microsoft IIS/OWA. Install Guide

1000Mbps Ethernet Performance Test Report

TEST METHODOLOGY. Endpoint Protection Evasion and Exploit. v4.0

By the Citrix Publications Department. Citrix Systems, Inc.

Dell InTrust Preparing for Auditing Cisco PIX Firewall

Dell One Identity Cloud Access Manager How to Configure vworkspace Integration

Dell One Identity Cloud Access Manager SonicWALL Integration Overview

Remote Firewall Deployment

Managing for the Long Term: Keys to Securing, Troubleshooting and Monitoring a Private Cloud

SolarWinds. Packet Analysis Sensor Deployment Guide

NEXT-GENERATION FIREWALL

Global Enterprise Network Firewall Market

Performance Evaluation of VMXNET3 Virtual Network Device VMware vsphere 4 build

Securing Amazon It s a Jungle Out There

Dell Spotlight on Active Directory Deployment Guide

z/os V1R11 Communications Server system management and monitoring

Dell Statistica Statistica Enterprise Installation Instructions

Performance Analysis and Software Optimization on Systems Using the LAN91C111

Application Notes for Configuring a SonicWALL VPN with an Avaya IP Telephony Infrastructure - Issue 1.0

Firewall Sandwich. Aleksander Kijewski Presales Engineer Dell Software Group. Dell Security Peak Performance

Adaptec: Snap Server NAS Performance Study

TEST METHODOLOGY. Secure Web Gateway (SWG) v1.5.1

Transcription:

FIREWALL COMPARATIVE ANALYSIS Performance 2013 Thomas Skybakmoen, Francisco Artes, Bob Walder, Ryan Liles Tested Products Barracuda F800, Check Point 12600, Cyberoam CR2500iNG, Dell SonicWALL NSA 4500, Fortinet FortiGate- 800c, Juniper SRX550, NETASQ NG1000- A, NETGEAR ProSecure UTM9S, Palo Alto Networks PA- 5020, Sophos UTM 425, Stonesoft FW- 1301, WatchGuard XTM 1050

Overview Implementation of a firewall can be a complex process with multiple factors affecting the overall performance of the solution. Each of these factors should be considered over the course of the useful life of the solution, including: 1. Deployment use cases: a. Will the Firewall be deployed to protect servers, desktop clients, or both? 2. What does the traffic look like? a. Concurrency and connection rates. b. Connections per second and capacity with different traffic profiles. c. Latency and application response times. There is usually a trade- off between security effectiveness and performance; a product s security effectiveness should be evaluated within the context of its performance (and vice versa). This ensures that new security protections do not adversely impact performance and security shortcuts are not taken to maintain or improve performance. Sizing considerations are absolutely critical, since vendor performance claims can vary significantly from actual throughput with protection turned on. Farther to the right indicates higher rated throughput. Higher up indicates more connections per second. Products with low connections/throughput ratio run the risk of running out of connections before they reach their maximum potential throughput. 250,000 Maximum TCP Connections per Second 200,000 150,000 100,000 50,000 - - 2,000 4,000 6,000 8,000 12,000 NSS Rated Throughput (Mbps) Figure 1: Throughput and Connection Rates 2013 NSS Labs, Inc. All rights reserved. 2

Table of Contents Overview... 2 Analysis... 4 UDP Throughput & Latency... 5 Connection Dynamics Concurrency and Connection Rates... 7 HTTP Connections per Second and Capacity... 9 HTTP Connections per Second and Capacity (Throughput)... 9 Application Average Response Time - HTTP (at 90% Max Capacity)... 12 Real- World Traffic Mixes... 13 Test Methodology... 14 Contact Information... 14 Table of Figures Figure 1: Throughput and Connection Rates... 2 Figure 2: Vendor Claimed vs NSS Rated Throughput in Mbps... 4 Figure 3: UDP Throughput by Packet Size (I)... 5 Figure 4: UDP Throughput by Packet Size (II)... 6 Figure 5: UDP Latency by Packet Size (Microseconds)... 6 Figure 6: Concurrency and Connection Rates (I)... 7 Figure 7: Concurrency and Connection Rates (II)... 8 Figure 8: Maximum Throughput Per Device With 44Kbyte Response... 9 Figure 9: Maximum Throughput Per Device With 21Kbyte Response... 10 Figure 10: Maximum Throughput Per Device With 10Kbyte Response... 10 Figure 11: Maximum Throughput Per Device With 4.5Kbyte Response... 11 Figure 12: Maximum Throughput Per Device With 1.7Kbyte Response... 11 Figure 13: Maximum Connection Rates Per Device With Various Response Sizes... 12 Figure 14: Application Latency (Microseconds) Per Device With Various Response Sizes... 12 Figure 15: Real- World Performance by Device... 13 2013 NSS Labs, Inc. All rights reserved. 3

Analysis NSS Labs research indicates that the majority of enterprises will deploy traditional firewalls in front of their datacenters and at the core of the network to separate unrelated traffic. Because of this, NSS rates product performance based upon the average of three traffic types: 21KB HTTP response traffic, a mix of perimeter traffic common in enterprises, and a mix of internal core traffic common in enterprises. Details of these traffic mixes are available in the Firewall Test Methodology (www.nsslabs.com). Every effort is made to deploy policies that ensure the optimal combination of security effectiveness and performance, as would be the aim of a typical customer deploying the device in a live network environment. This provides readers with the most useful information on key firewall security effectiveness and performance capabilities based upon their expected usage. NSS Throughput Rating Vendor Stated Throughput - 5,000 15,000 20,000 25,000 30,000 7,827 9,200 8,400 8,733 28,000 850 990 9,667 20,000 2,127 5,500 2,540 7,000 231 850 4,120 5,000 3,000 5,147 5,000 6,000 2,200 Figure 2: Vendor Claimed vs. NSS Rated Throughput in Mbps The results presented in the chart above show the difference between the NSS performance rating and vendor performance claims, which are often under ideal/unrealistic conditions. Where multiple figures are quoted by vendors, NSS selects those that relate to TCP, or with protection enabled, performance expectations, rather than the more optimistic UDP- only or large packet size performance figures often quoted. Even so, NSS rated throughput is typically lower than that claimed by the vendor, often significantly so, since it is more representative of how devices will perform in real- world deployments. 2013 NSS Labs, Inc. All rights reserved. 4

UDP Throughput & Latency The aim of this test is to determine the raw packet processing capability of each in- line port pair of the device only. This traffic does not attempt to simulate any form of real- world network condition. No TCP sessions are created during this test, and there is very little for the detection engine to do in the way of protocol analysis. However, this test is relevant since vendors are forced to perform inspection on UDP packets as a result of VoIP, video, and other streaming applications. 20,000 18,000 16,000 14,000 Megabits per Second 12,000 8,000 6,000 4,000 2,000-64 Byte Packets 128 Byte Packets2 256 Byte Packets 512 Byte Packets 1024 Byte Packets 1514 Byte Packets UDP Packet Size Figure 3: UDP Throughput by Packet Size (I) Fortinet s Fortigate 800c was the only device to demonstrate anything close to line rate capacity with packet sizes from 1514 bytes all the way down to 64 bytes. In addition, it was the only device to consistently demonstrate latency of less than 10 microseconds. The Juniper SRX 550 demonstrated the second best latency at 12-16 microseconds. 2013 NSS Labs, Inc. All rights reserved. 5

Product 64 Byte Packets 128 Byte Packets 256 Byte Packets 512 Byte Packets 1024 Byte Packets 1514 Byte Packets Barracuda F800 780 2,000 3,700 6,500 12,600 18,500 Check Point 12600 1,900 3,400 6,400 11,700 12,100 12,200 Cyberoam CR2500iNG 3,300 9,300 9,300 11,230 16,200 19,300 Dell SonicWALL NSA 4500 120 400 400 890 1,700 2,600 Fortinet FortiGate- 800c 18,900 19,500 19,600 19,700 19,800 19,900 Juniper SRX550 400 1,450 1,450 2,800 5,500 8,000 NETASQ NG1000- A 350 600 1,200 2,300 4,800 7,300 NETGEAR ProSecure UTM9S 2 12 12 20 64 102 Palo Alto Networks PA- 5020 5,092 9,400 9,400 9,597 9,785 9,935 Sophos UTM 425 640 2,450 2,450 4,600 6,000 6,000 Stonesoft FW- 1301 2,600 4,400 8,300 14,900 19,100 19,900 WatchGuard XTM 1050 346 925 925 2,000 4,100 6,700 Figure 4: UDP Throughput by Packet Size (II) In- line security devices that introduce high levels of latency are unacceptable, especially where multiple security devices are placed in the data path. The chart below reflects the latency (in microseconds) as recorded during the UDP throughput tests at 90% of maximum load. Lower values are preferred. Product 64 Byte Packets - Latency (μs) 128 Byte Packets - Latency (μs) 256 Byte Packets - Latency (μs) 512 Byte Packets - Latency (μs) 1024 Byte Packets - Latency (μs) 1514 Byte Packets - Latency (μs) Barracuda F800 273 163 108 104 103 109 Check Point 12600 75 124 99 82 102 109 Cyberoam CR2500iNG 1,185 845 452 385 302 270 Dell SonicWALL NSA 4500 30 31 32 33 37 42 Fortinet FortiGate- 800c 5 6 6 7 8 9 Juniper SRX550 12 12 12 13 14 16 NETASQ NG1000- A 36 36 43 46 47 36 NETGEAR ProSecure UTM9S 232 237 243 255 337 603 Palo Alto Networks PA- 5020 15 19 22 26 33 38 Sophos UTM 425 59 61 60 63 92 169 Stonesoft FW- 1301 50 84 51 54 82 81 WatchGuard XTM 1050 136 156 182 269 460 705 Figure 5: UDP Latency by Packet Size (Microseconds) 2013 NSS Labs, Inc. All rights reserved. 6

Connection Dynamics Concurrency and Connection Rates These tests stress the detection engine to determine how the sensor copes with increasing rates of TCP connections per second, application layer transactions per second, and concurrent open connections. All packets contain valid payload and address data and these tests provide an excellent representation of a live network at various connection/transaction rates. Note that in all tests, the following critical breaking points where the final measurements are taken are used: Excessive concurrent TCP connections - latency within the firewall is causing unacceptable increase in open connections on the server- side. Excessive response time for HTTP transactions/smtp sessions - latency within the firewall is causing excessive delays and increased response time to the client. Unsuccessful HTTP transactions/smtp sessions normally, there should be zero unsuccessful transactions. Once these appear, it is an indication that excessive latency within the firewall is causing connections to time out. The following are the key connection dynamics results from the performance tests. Product Max. Concurrent TCP Connections Max. Concurrent TCP Connections w/data TCP Connections Per Second HTTP Connections Per Second HTTP Transactions Per Second Barracuda F800 1,000,000 1,000,000 134,200 117,000 356,000 Check Point 12600 1,500,000 1,500,000 52,000 113,000 391,700 Cyberoam CR2500iNG 3,100,000 2,999,000 215,000 179,000 360,000 Dell SonicWALL NSA 4500 400,000 400,000 15,600 10,300 25,000 Fortinet FortiGate- 800c 6,000,000 4,400,000 180,000 180,000 397,000 Juniper SRX550 533,000 542,000 18,500 18,000 80,000 NETASQ ng1000- A 1,280,000 1,200,000 57,000 43,500 102,300 NETGEAR ProSecure UTM9S 16,700 15,000 760 480 6,400 Palo Alto Networks PA- 5020 1,000,000 1,000,000 36,000 36,000 348,000 Sophos UTM 425 480,000 2,000,000 12,000 31,800 270,000 Stonesoft FW- 1301 6,000,000 2,900,000 57,000 51,000 328,600 WatchGuard XTM 1050 2,600,000 2,600,000 19,000 18,000 136,000 Figure 6: Concurrency and Connection Rates (I) 2013 NSS Labs, Inc. All rights reserved. 7

Beyond overall throughput of the device, connection dynamics can play an important role in sizing a security device that will not unduly impede the performance of a system or an application. Maximum connection and transaction rates help size a device more accurately than simply focusing on throughput. By knowing the maximum connections per second, it is possible to predict maximum throughput based upon the traffic mix in a given enterprise environment. For example If the device maximum HTTP CPS is 2,000, and average traffic size is 44KB such that 2,500 CPS = 1Gbps, then the tested device will achieve a maximum of 800 Mbps ((2,000/2,500) x 1,000 Mbps)) = 800 Mbps. Maximum concurrent TCP connections and maximum TCP connections per second rates are also useful metrics when attempting to size a device accurately. Products with low connection/throughput ratio run the risk of exhausting connections before they reach their maximum potential throughput. By knowing the maximum connections per second, it is possible to predict when a device will fail in a given enterprise environment. 250,000 200,000 Maximum TCP Connections per Second 150,000 100,000 50,000 - - 1,000,000 2,000,000 3,000,000 4,000,000 5,000,000 6,000,000 7,000,000 Maximum Concurrent / Simultaneous TCP Connections Figure 7: Concurrency and Connection Rates (II) Higher up indicates increased connections per second capacity. Farther to the right indicates increased concurrent / simultaneous connections. Products with low concurrent connection / connection per second ratio run the risk of exhausting connections (sessions) before they reach their maximum potential connection rate. 2013 NSS Labs, Inc. All rights reserved. 8

HTTP Connections per Second and Capacity In- line firewall devices exhibit an inverse correlation between security effectiveness and performance. The more deep- packet inspection is performed, the fewer packets can be forwarded. Furthermore, it is important to consider a real- world mix of traffic that a device will encounter. NSS tests aim to stress the HTTP detection engine in order to determine how the sensor copes with detecting and blocking under network loads of varying average packet size and varying connections per second. By creating genuine session- based traffic with varying session lengths, the sensor is forced to track valid TCP sessions, thus ensuring a higher workload than for simple packet- based background traffic. Each transaction consists of a single HTTP GET request and there are no transaction delays (i.e. the web server responds immediately to all requests). All packets contain valid payload (a mix of binary and ASCII objects) and address data. This test provides an excellent representation of a live network (albeit one biased towards HTTP traffic) at various network loads. HTTP Connections per Second and Capacity (Throughput) As previously stated, NSS research has found that there is usually a trade- off between security effectiveness and performance. Because of this, it is important to judge a product s security effectiveness within the context of its performance (and vice versa). This ensures that new security protections do not adversely impact performance and security shortcuts are not taken to maintain or improve performance. The following charts compare maximum connection rate (HTTP CPS), maximum rated throughput (Mbps) and average application latency (average HTTP response time in milliseconds) across a range of HTTP response sizes. Megabits per Second - 1,000 2,000 3,000 4,000 5,000 6,000 7,000 8,000 9,000 960 2,400 3,080 209 4,120 3,000 3,320 2,520 WatchGuar d XTM 1050 Stonesoft FW-1301 Sophos UTM 425 Palo Alto Networks PA-5020 NETGEAR ProSecure UTM9S NETASQ NG1000-A Juniper SRX550 Fortinet FortiGate-8 00c Dell SonicWALL NSA 4500 Cyberoam CR2500iN G Check Point 12600 Barracuda F800 44Kbyte Response 2,520 3,320 3,000 4,120 209 3,080 2,400 960 Figure 8: Maximum Throughput Per Device With 44Kbyte Response 2013 NSS Labs, Inc. All rights reserved. 9

Megabits per Second - 1,000 2,000 3,000 4,000 5,000 6,000 7,000 8,000 9,000 8,780 820 1,880 2,720 112 4,160 3,000 2,940 2,200 WatchGuar d XTM 1050 Stonesoft FW-1301 Sophos UTM 425 Palo Alto Networks PA-5020 NETGEAR ProSecure UTM9S NETASQ NG1000-A Juniper SRX550 Fortinet FortiGate-8 00c Dell SonicWAL L NSA 4500 Cyberoam CR2500iN G 21Kbyte Response (Mbps) 2,200 2,940 3,000 4,160 112 2,720 1,880 820 8,780 Check Point 12600 Barracuda F800 Figure 9: Maximum Throughput Per Device With 21Kbyte Response Megabits per Second - 1,000 2,000 3,000 4,000 5,000 6,000 7,000 8,000 9,000 6,490 7,460 8,491 520 7,700 1,260 2,010 58 3,280 3,000 2,910 1,610 WatchGuard XTM 1050 Stonesoft FW-1301 Sophos UTM 425 Palo Alto Networks PA-5020 NETGEAR ProSecure UTM9S NETASQ NG1000-A Juniper SRX550 Fortinet FortiGate-800 c Dell SonicWALL NSA 4500 Cyberoam CR2500iNG Check Point 12600 10KB Response (Mbps) 1,610 2,910 3,000 3,280 58 2,010 1,260 7,700 520 8,491 7,460 6,490 Barracuda F800 Figure 10: Maximum Throughput Per Device With 10Kbyte Response 2013 NSS Labs, Inc. All rights reserved. 10

Megabits per Second - 1,000 2,000 3,000 4,000 5,000 6,000 7,000 8,000 9,000 4,390 4,660 5,650 280 6,850 770 1,360 32 1,680 2,600 2,200 965 WatchGuard XTM 1050 Stonesoft FW-1301 Sophos UTM 425 Palo Alto Networks PA-5020 NETGEAR ProSecure UTM9S NETASQ NG1000-A Juniper SRX550 Fortinet FortiGate-800c Dell SonicWALL NSA 4500 Cyberoam CR2500iNG Check Point 12600 4.5KB Response (Mbps) 965 2,200 2,600 1,680 32 1,360 770 6,850 280 5,650 4,660 4,390 Barracuda F800 Figure 11: Maximum Throughput Per Device With 4.5Kbyte Response Megabits per Second - 1,000 2,000 3,000 4,000 5,000 6,000 7,000 8,000 9,000 2,603 2,710 3,675 240 3,925 440 920 17 1,150 1,400 1,050 525 WatchGuard XTM 1050 Stonesoft FW-1301 Sophos UTM 425 Palo Alto Networks PA-5020 NETGEAR ProSecure UTM9S NETASQ NG1000-A Juniper SRX550 Fortinet FortiGate-800c Dell SonicWALL NSA 4500 Cyberoam CR2500iNG Check Point 12600 1.7KB Response (Mbps) 525 1,050 1,400 1,150 17 920 440 3,925 240 3,675 2,710 2,603 Barracuda F800 Figure 12: Maximum Throughput Per Device With 1.7Kbyte Response 2013 NSS Labs, Inc. All rights reserved. 11

The following table shows the number of HTTP connections per second required to achieve the rated throughput. Product 44Kbyte Response 21Kbyte Response 10Kbyte Response 4.5Kbyte Response 1.7Kbyte Response Barracuda F800 25,000 43,900 64,900 87,800 104,100 Check Point 12600 25,000 50,000 74,600 93,200 108,400 Cyberoam CR2500iNG 25,000 50,000 84,908 113,000 147,000 Dell SonicWALL NSA 4500 2,400 4,100 5,200 5,600 9,600 Fortinet FortiGate- 800c 25,000 50,000 77,000 137,000 157,000 Juniper SRX550 6,000 9,400 12,600 15,400 17,600 NETASQ ng1000- A 7,700 13,600 20,100 27,200 36,800 NETGEAR ProSecure UTM9S 522 562 582 642 661 Palo Alto Networks PA- 5020 10,300 20,800 32,800 33,600 46,000 Sophos UTM 425 7,500 15,000 30,000 52,000 56,000 Stonesoft FW- 1301 8,300 14,700 29,100 44,000 42,000 WatchGuard XTM 1050 6,300 11,000 16,100 19,300 21,000 Figure 13: Maximum Connection Rates Per Device With Various Response Sizes Application Average Response Time - HTTP (at 90% Max Capacity) The following table details the average application response time (application latency) with various traffic sizes at 90% Max Capacity (Throughput). The lower the number the better (improved application response time). Juniper SRX 3600 demonstrated the lowest (best) application response times, while the NETGEAR ProSecure UTM9S introduced the most application latency. Product 44Kbyte Latency (ms) 21Kbyte Latency (ms) 10Kbyte Latency (ms) 4.5Kbyte Latency (ms) 1.7Kbyte Latency (ms) Barracuda F800 1.7 1.1 0.9 0.3 0.3 Check Point 12600 1.81 1.08 1.1 0.35 0.28 Cyberoam CR2500iNG 1.61 0.83 0.7 0.22 0.24 Dell SonicWALL NSA 4500 1.99 1.02 0.04 0.04 0.07 Fortinet FortiGate- 800c 0.94 0.78 0.53 0.4 0.34 Juniper SRX550 0.14 0.04 0.04 0.01 0.03 NETASQ ng1000- A 1.42 1.25 0.53 0.12 0.16 NETGEAR ProSecure UTM9S 2.55 1.39 2.43 0.93 0.65 Palo Alto Networks PA- 5020 1.01 0.63 0.27 0.1 0.08 Sophos UTM 425 1.98 0.99 0.75 0.06 0.03 Stonesoft FW- 1301 1.1 0.9 0.4 0.2 0.08 WatchGuard XTM 1050 5.1 3.01 2.14 0.85 0.77 Figure 14: Application Latency (Microseconds) Per Device With Various Response Sizes 2013 NSS Labs, Inc. All rights reserved. 12

Real- World Traffic Mixes The aim of these tests is to measure the performance of the device under test (DUT) in a real world environment by introducing additional protocols and real content, while still maintaining a precisely repeatable and consistent background traffic load. In order to simulate real use cases, different protocol mixes are utilized to model placement of the DUT within various locations on a corporate network. For details about real world traffic protocol types and percentages, see the NSS Labs Firewall Test Methodology, available at www.nsslabs.com. Real World Protocol Mix (Perimeter) Real World Protocol Mix (Core) - 2,000 4,000 6,000 8,000 4,700 5,200 6,200 950 780 9,000 1,500 1,200 3,000 3,700 121 460 4,500 3,700 3,000 3,000 3,800 2,600 1,800 8,700 Figure 15: Real- World Performance by Device Most vendors perform better in the real- world protocol mix (perimeter,) a protocol mix typically seen at an enterprise perimeter. However Fortinet Fortigate- 800c is the only vendor to scale equally well in the real- world protocol mix (core), a protocol mix typical of that seen in a large datacenter or the core of an enterprise network. 2013 NSS Labs, Inc. All rights reserved. 13

Test Methodology Methodology Version: Firewall v4 A copy of the test methodology is available on the NSS Labs website at www.nsslabs.com Contact Information NSS Labs, Inc. 206 Wild Basin Rd, Suite 200A Austin, TX 78746 USA +1 (512) 961-5300 info@nsslabs.com www.nsslabs.com v2013.02.07 This and other related documents available at: www.nsslabs.com. To receive a licensed copy or report misuse, please contact NSS Labs at +1 (512) 961-5300 or sales@nsslabs.com. 2013 NSS Labs, Inc. All rights reserved. No part of this publication may be reproduced, photocopied, stored on a retrieval system, or transmitted without the express written consent of the authors. Please note that access to or use of this report is conditioned on the following: 1. The information in this report is subject to change by NSS Labs without notice. 2. The information in this report is believed by NSS Labs to be accurate and reliable at the time of publication, but is not guaranteed. All use of and reliance on this report are at the reader s sole risk. NSS Labs is not liable or responsible for any damages, losses, or expenses arising from any error or omission in this report. 3. NO WARRANTIES, EXPRESS OR IMPLIED ARE GIVEN BY NSS LABS. ALL IMPLIED WARRANTIES, INCLUDING IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, AND NON- INFRINGEMENT ARE DISCLAIMED AND EXCLUDED BY NSS LABS. IN NO EVENT SHALL NSS LABS BE LIABLE FOR ANY CONSEQUENTIAL, INCIDENTAL OR INDIRECT DAMAGES, OR FOR ANY LOSS OF PROFIT, REVENUE, DATA, COMPUTER PROGRAMS, OR OTHER ASSETS, EVEN IF ADVISED OF THE POSSIBILITY THEREOF. 4. This report does not constitute an endorsement, recommendation, or guarantee of any of the products (hardware or software) tested or the hardware and software used in testing the products. The testing does not guarantee that there are no errors or defects in the products or that the products will meet the reader s expectations, requirements, needs, or specifications, or that they will operate without interruption. 5. This report does not imply any endorsement, sponsorship, affiliation, or verification by or with any organizations mentioned in this report. 6. All trademarks, service marks, and trade names used in this report are the trademarks, service marks, and trade names of their respective owners. 2013 NSS Labs, Inc. All rights reserved. 14