The Advantages of Using EIGRP on an Enterprise Network



Similar documents
Top-Down Network Design

A REPORT ON ANALYSIS OF OSPF ROUTING PROTOCOL NORTH CAROLINA STATE UNIVERSITY

Table of Contents. Cisco How Does Load Balancing Work?

Cisco Discovery 3: Introducing Routing and Switching in the Enterprise hours teaching time

SUNYIT. Reaction Paper 2. Measuring the performance of VoIP over Wireless LAN

Internet Firewall CSIS Packet Filtering. Internet Firewall. Examples. Spring 2011 CSIS net15 1. Routers can implement packet filtering

Performance Analysis of Routing Protocols for Real Time Application

ISOM3380 Advanced Network Management. Spring Course Description

CompTIA Exam N CompTIA Network+ certification Version: 5.1 [ Total Questions: 1146 ]

Transport and Network Layer

Security Labs in OPNET IT Guru

Quality of Service Analysis of Video Conferencing over WiFi and Ethernet Networks

Disaster Recovery Design Ehab Ashary University of Colorado at Colorado Springs

SSVP SIP School VoIP Professional Certification

Virtualized Open-Source Network Security Appliance

The WestNet Advantage: -- Textbooks, ebooks, ecourses -- Instructor Resourse Center -- Student Resource Center

Enterprise Edge Communications Manager. Data Capabilities

Best Effort gets Better with MPLS. Superior network flexibility and resiliency at a lower cost with support for voice, video and future applications

This chapter covers four comprehensive scenarios that draw on several design topics covered in this book:

Usage of OPNET IT tool to Simulate and Test the Security of Cloud under varying Firewall conditions

Performance Comparison of Mixed Protocols Based on EIGRP, IS-IS and OSPF for Real-time Applications

Technical Glossary from Frontier

Broadband Phone Gateway BPG510 Technical Users Guide

Performance of voice and video conferencing over ATM and Gigabit Ethernet backbone networks

November Defining the Value of MPLS VPNs

A host-based firewall can be used in addition to a network-based firewall to provide multiple layers of protection.

UPPER LAYER SWITCHING

Owner of the content within this article is Written by Marc Grote

Router and Routing Basics

UIP1868P User Interface Guide

Chapter 3. Enterprise Campus Network Design

Lesson 5-3: Border Gateway Protocol

CHAPTER 6 DESIGNING A NETWORK TOPOLOGY

COURSE AGENDA. Lessons - CCNA. CCNA & CCNP - Online Course Agenda. Lesson 1: Internetworking. Lesson 2: Fundamentals of Networking

NComputing L-Series LAN Deployment

Dante: Know It, Use It, Troubleshoot It.

Chapter 4. Distance Vector Routing Protocols

Lab Diagramming External Traffic Flows

ICTTEN6172A Design and configure an IP- MPLS network with virtual private network tunnelling

Secured Voice over VPN Tunnel and QoS. Feature Paper

MPLS/BGP Network Simulation Techniques for Business Enterprise Networks

VPLS lies at the heart of our Next Generation Network approach to creating converged, simplified WANs.

Cisco Configuring Basic MPLS Using OSPF

Home Networking Evaluating Internet Connection Choices for a Small Home PC Network

SBSCET, Firozpur (Punjab), India

Data Networking and Architecture. Delegates should have some basic knowledge of Internet Protocol and Data Networking principles.

Comparison of RIP, EIGRP, OSPF, IGRP Routing Protocols in Wireless Local Area Network (WLAN) By Using OPNET Simulator Tool - A Practical Approach

NETWORK DESIGN BY USING OPNET IT GURU ACADEMIC EDITION SOFTWARE

The Analysis and Simulation of VoIP

BSCI Chapter Cisco Systems, Inc. All rights reserved.

ICTTEN4215A Install and configure internet protocol TV in a service provider network

MPLS: Key Factors to Consider When Selecting Your MPLS Provider Whitepaper

GR2000: a Gigabit Router for a Guaranteed Network

Cisco Dynamic Multipoint VPN: Simple and Secure Branch-to-Branch Communications

Associate in Science Degree in Computer Network Systems Engineering

Networking Topology For Your System

Planeamento e Gestão de Redes. Análise de Requisitos

SSVVP SIP School VVoIP Professional Certification

BGP: Border Gateway Protocol

WHY CHOOSE COX BUSINESS FOR YOUR COMPANY S NETWORK SERVICE NEEDS?

Voice Over IP. MultiFlow IP Phone # 3071 Subnet # Subnet Mask IP address Telephone.

MITEL. NetSolutions. Flat Rate MPLS VPN

IP Telephony Management

Enhancing Network Monitoring with Route Analytics

WIDE AREA NETWORKS: A COMPARISON OF DIFFERENT APPROACHES FOR HANDS-ON LABS

Lab 3: Evaluating Application Performance across a WAN

WHITEPAPER MPLS: Key Factors to Consider When Selecting Your MPLS Provider

Troubleshooting and Maintaining Cisco IP Networks Volume 1

: Interconnecting Cisco Networking Devices Part 2 v1.1

Deploying in a Distributed Environment

CORPORATE NETWORKING

Testing Network Security Using OPNET

Avaya ExpertNet Lite Assessment Tool

Implementing, Managing, and Maintaining a Microsoft Windows Server 2003 Network Infrastructure

Global Learning & Development CCNA Curriculum Overview

Networking 4 Voice and Video over IP (VVoIP)

Distance Vector Routing Protocols. Routing Protocols and Concepts Ola Lundh

The OSI Model: Understanding the Seven Layers of Computer Networks

Lab 1: Evaluating Internet Connection Choices for a Small Home PC Network

Sprint Global MPLS VPN IP Whitepaper

IP Routing Configuring Static Routes

MPLS Traffic Engineering in ISP Network

1.1. Abstract VPN Overview

White Paper Copyright 2011 Nomadix, Inc. All Rights Reserved. Thursday, January 05, 2012

SCOPE DOCUMENT. Trade Name IT- Network Systems Administration Post- Secondary DATE OF DISTRIBUTION VIA WEBSITE

Optimizing Enterprise Network Bandwidth For Security Applications. Improving Performance Using Antaira s Management Features

Hosting more than one FortiOS instance on. VLANs. 1. Network topology

Avaya IP Office. Converged Communications. Contact Centres Unified Communication Services

How To Design A Network For A Small Business

BGP. 1. Internet Routing

Inquire about our programs at Worcester Technical High School!

Introduction to Routing and Packet Forwarding. Routing Protocols and Concepts Chapter 1

Efficient Video Distribution Networks with.multicast: IGMP Querier and PIM-DM

Layer 3 Network + Dedicated Internet Connectivity

Optimization of Computer Network for Efficient Performance

Enhance student s learning with an aid of simulation software to understand Computer Networking Undergraduate courses.

Routing with OSPF. Introduction

Interconnecting Cisco Networking Devices Part 2

Zarząd (7 osób) F inanse (13 osób) M arketing (7 osób) S przedaż (16 osób) K adry (15 osób)

Transcription:

The Advantages of Using EIGRP on an Enterprise Network By Randy Patrick Submitted to the Faculty of the Information Technology Program in Partial Fulfillment of the Requirements for the Degree of Bachelor of Science in Information Technology University of Cincinnati College of Applied Science May 2007

The Advantages of Using EIGRP on an Enterprise Network by Randy Patrick Submitted to the Faculty of the Information Technology Program in Partial Fulfillment of the Requirements for the Degree of Bachelor of Science in Information Technology Copyright 2007 Randy Patrick The author grants to the Information Technology Program permission to reproduce and distribute copies of this document in whole or in part. Randy Patrick Professor John Nyland Hazem Said, Ph.D. Department Head Date Date Date 2

Table of Contents Section Page Table of Contents 3 List of Figures 4 Abstract 6 1. Statement of the Problem 7 2. Description of the Solution 9 2.1 User Profiles 10 2.2 Design Protocol 11 3. Deliverables 13 4. Design and Development 14 4.1 Budget 14 4.2 Timeline 15 4.3 Software 16 4.4 Hardware 17 5. Proof of Design 17 5.1 Scenarios 17 5.2 Complete Virtual Network 20 5.3 Convergence Times 21 5.4 Logical Layout 25 5.5 Application Delay Times 27 5.6 Overall bit flow 31 5.7 Results with a Failed Node 34 5.8 VOIP Traffic 36 6. Conclusions and Recommendations 41 6.1 Conclusions 42 6.2 Recommendations 42 3

List of Figures Figure Number Page Figure 1. Network Overview 10 Figure 2. Physical Layout 12 Figure 3. Project Budget 15 Figure 4. Project Timeline 16 Figure 5. Campus Backbone 18 Figure 6. Edge Distribution 18 Figure 7. Campus Backbone with Failure 19 Figure 8. Edge Distribution with Failure 20 Figure 9. Complete Virtual Network 21 Figure 10. Convergence time Campus Backbone General Traffic Heavy Load 22 Figure 11. Convergence time - Campus Backbone with Failure General Traffic 23 Heavy Load Figure 12. Convergence time Edge Distribution - General Traffic - Heavy Load 24 Figure 13. Convergence time - Edge Distribution with Failure General Traffic 25 Heavy Load Figure 14. IP Address Layouts 26 Figure 15. Database Entry response time Campus Backbone Heavy Load 27 Figure 16. FTP Download response time Campus Backbone Heavy Load 28 Figure 17. Database Entry response time Campus Backbone - Light Load 29 Figure 18. Database Entry response time Edge Distribution - Heavy Load 30 Figure 19. Database Entry response time Edge Distribution - Light Load 31 4

Figure 20. Overall Bit Flow Campus Backbone - General Traffic Heavy Load 32 Figure 21. Overall Bit Flow Campus Backbone 33 Video Conferencing Traffic Only Figure 22. Overall Bit Flow Edge Distribution General Traffic Heavy Load 34 Figure 23. Results with a Failed Node Database Entry traffic sent 35 Campus Backbone Heavy Load Figure 24. Edge Distribution with Failure - Database Entry traffic sent - 36 General traffic - Heavy Load Figure 25. VOIP packet delay variation Campus Backbone 37 Figure 26. VOIP packet end-to-end delay Campus Backbone 38 Figure 27. VOIP Packet end-to-end delay Edge Distribution 39 Figure 28. VOIP Packet Delay Variation Edge Distribution with failure 40 Heavy Load Figure 29. VOIP Packet Delay Variation Edge Distribution with failure - 40 Light Load Figure 30. VOIP traffic only Edge Distribution 41 5

Abstract This project shows the advantages of using the Enhanced Interior Gateway Routing Protocol (EIGRP) on an enterprise network. To show the advantages of using EIGRP on an enterprise network, a virtual network was created based on Cisco s Enterprise Composite Network Model. With this network, tests were completed to show how the network traffic flows for EIGRP. These tests were also compared with another common routing protocol, the Routing Information Protocol. The tests were completed to show the different convergence times, delay times, packet flow, and bandwidth usage. The virtual network was divided into sections, and the tests were completed on each section of the network for a specific amount of simulated time. The benefit of the research provided by this project will show the stability and consistency of using EIGRP on an enterprise network 6

The Advantages of using EIGRP on an Enterprise Network 1. Statement of the Problem Many routing protocols are used in enterprise networking. The most popular routing protocols include the Routing Information Protocol (RIP), RIP v2, Interior Gateway Routing Protocol (IGRP), Open Shortest Path First (OSPF), and EIGRP. EIGRP is the latest routing protocol of the group, and it is often seen as the most efficient and most robust routing protocol in use (3). The differences between the routing protocols RIP and EIGRP are vast. RIP takes some algorithms that were developed in 1957, and uses them in a way to read the distances on a network. RIP is still considered one of the most commonly used routing protocols on internal networks today (3). The limitations of RIP, however, stand out when it is compared with EIGRP. The biggest limitation of RIP is the maximum hop count of 15. If a hop count reaches 16 hops, then that means that the 16 th hop is seen as a loop to infinity. This is a measure taken to reduce network convergence time, and it also limits the size of the network. What this limited hop count means is that there cannot be more than 15 hops from one individual router or switch. This hop system is known as a metric system and the lower the metric, the better (2). EIGRP was introduced as an improvement over IGRP. Considering IGRP was Cisco s improvement over RIP, EIGRP should be an obvious choice for most enterprise networks (5). Some networks still use the RIP routing protocol, and it is clear that there are many advantages for those networks to use EIGRP. 7

To show the advantages of using EIGRP on an enterprise network, a network was created that was able to support both RIP and EIGRP. For this network, a physical and logical layout was necessary. The physical layout provides an easy to read overview of the network. This physical layout was created with Microsoft Visio Professional 2003. This application is very useful for creating network layouts. Symbols and designs were used to make the network easier to understand, and clarify what was done. The logical layout was provided for all of the IP addresses for the network. Also, the proper subnetting was necessary, and it was confirmed with a professional subnetting calculator. The logical layout is represented in a table with the IP Addresses for each device listed. For this project to be complete, IP Addresses must be assigned on the network. This virtual network was based on the Cisco Composite Enterprise Network Model. This network model sets an example of a solid network by showing where certain network functions should be placed and implemented (13). Using a common network foundation such as the Composite Enterprise Network Model provided practical uses for this research. This project will show the advantages of using the Enhanced Interior Gateway Routing Protocol (EIGRP) on an Enterprise network. The problem is, however, finding how to show these advantages. Since the University of Cincinnati has the resources to use virtual networking software, it was best to take this route to prove the advantages of using EIGRP on an enterprise network. 8

2. Description of the Solution The purpose of this project is to show the advantages of using the Enhanced Interior Gateway Routing Protocol (EIGRP) on an enterprise network, and the research from this project shows those advantages. To show the Advantages of using EIGRP on an enterprise network, it is necessary to use a virtual network based on Cisco s Enterprise Composite Network Model (ECNM). The ECNM is a model network that many large companies use as a template for their network infrastructure. This virtual network was created in the OPNET IT Guru virtual network software. The virtual network is composed of several key sections. At the top of the network is a server farm connected to the campus backbone. Under the campus backbone, the three building distributions exist. From the campus backbone, the edge distribution switches are setup. From the edge distribution, four Demilitarized Zones (DMZ) are created for various secure connections across the network. Ultimately, a DMZ makes sure that traffic coming from outside of the network is secured and filtered so that it cannot do any damage to the internal network. Refer to Figure 1 for a better understanding of how this network looks. The Network Overview shows the server farm, campus backbone, and the building distributions. On the right side, there is the edge distribution and the four DMZs. 9

Figure 1 Network Overview To show the advantages of using EIGRP on this network, it is necessary to divide the network into sections. Once the network is divided into sections, the network must be set up with the necessary routing protocols. For this project, EIGRP will be compared directly with the Routing Information Protocol (RIP). These two routing protocols will be set up to run on specific sections of the network. When the correct application load is placed on the network, the tests can then run on these network sections. After the tests run on the different sections on this network, the results will then need to be analyzed and compared for a final solid conclusion. 2.1 User Profiles Only one user is necessary for this project. This user is the OPNET IT Guru Administrator, and the user profile is necessary because this user profile will run the tests in the OPNET software. With this profile, the network was created and the data will be 10

gathered. Since this OPNET software is being used for research, no other user profile will be necessary. 2.2 Design Protocol The first part of the Design Protocol was to actually create the virtual network. This virtual network was based on Cisco s design for the Enterprise Composite Network Model. Before building the virtual network in OPNET, it was necessary to create a physical layout in Visio. Refer to Figure 2 to see the physical layout for this project. The physical layout shows the way the devices are positioned in OPNET. After the physical layout was created, it was then approved by the technical advisor for this project. This approval was also for the specific Cisco devices that are being used in this virtual network. Once approval was obtained for the physical layout, the model was built in OPNET IT Guru. 11

Figure 2 Physical Layout The next step in the Design Protocol was to divide the network into sections that can be tested alone. Testing these sections alone, gave a better analysis for how the network functions as a whole. The network was divided into 2 main sections with different application load flowing over the network. First, the campus backbone which consists of the server farm, the server farm switches, and the campus backbone switches. The server farm switches are Cisco 2900 XL switches, and the Campus Backbone switches are Cisco 6509s. Next, the building distributions were set up, and they are same for three separate buildings. The building distribution routers are Cisco 12008 Gigabit switch routers, and the individual floor switches are Cisco 5500s. The building distributions were tested with the Campus Backbone. 12

The next section of the network to be tested is the edge distribution with all the DMZs. This section includes the edge distribution switches, and the four DMZs. The Edge Distribution switches are Cisco 6509s, and the internal routers to the DMZs are Cisco 12008 Gigabit switch routers. Also, the local switches in the DMZs are Cisco 2900 XL switches. It is also important to note that the tests that were completed on this network were not limited to the two main scenarios of Campus Backbone and Edge Distribution. Each of those scenarios was set up with different points of failure to see how the network responded with RIP and EIGRP. For example, the Campus Backbone had a simulation that was completed with a Campus Backbone Switch failed, and three building distribution routers failed. These results will be discussed later in this paper. After the network was divided into sections, the protocols will need to be set up on the network. There is an option in OPNET to enable RIP and EIGRP on the network. Besides setting up the routing protocols on each section of the network, the application load has to be applied to the network. The application load includes three types of traffic that were tested individually. First, the general traffic was simulated for each section of the network. The general traffic consists of Email, Web Browsing (HTTP), Database, File Transfers (FTP), and Print traffic. The other two traffics that are applied separately are voice and video traffic. The voice and video traffic were applied separately because of the length of a single simulation with that traffic. A single simulation will take about ten minutes with the Voice over IP (VOIP) or video conferencing traffic. It was also necessary to run the tests at different application load levels to see how consistent the routing protocols are over high load and low load. These application loads 13

are considered standards by Cisco from data that they have received over the last twentyfive years (9). Running high load and low load tests will provide a more thorough analysis to see if there is a difference between how the network responds to RIP and EIGRP. 3. Deliverables For a thorough analysis to be completed on this project, there were certain things things that needed to be completed and observed. During the design, testing, and analysis phases of this project, the following objectives were completed and observed: Complete Virtual Network Convergence times Logical layout Application delay times Overall bit flow Results with a failed node VOIP traffic 4. Design and Development 4.1 Budget The next sections discuss the project budget, timeline, software, and hardware. This project s budget did not change drastically over the year. The only change to the budget was the addition of a 60 GB external hard drive for storage use. That brought the budget total to $50,568.99. The actual cost for this project remains at zero dollars 14

because the hard drive was already purchased and the software was free. See Figure 3 for this project s budget. Item Subnetting Calculator 60 GB Hard Drive Description Provided by SolarWinds Have Retail Cost $0.00 69.99 Actual Cost $0.00 0.00 Microsoft Visio Professional 2003 OPNET IT Guru Provided by UC Provided by OPNET Retail Total: 499.00 50,000.00 $50,568.99 0.00 0.00 My Total: $0.00 Figure 3 Project Budget 4.2 Timeline The timeline for this project changed slightly over the year, but building the network and completing the testing definitely took the longest amount of time. View Figure 4 for the timeline of this project. 15

Figure 4 Project Timeline 4.3 Software Only three different types of software were necessary for this project. The main software being used for this project is called OPNET IT Guru. This networking software allows many different aspects of a networking to be viewed. Different types of networks can be built, and network protocols, connections, and devices can be analyzed. The second software used for this project is Microsoft Visio Professional 2003. This software was used to create the physical layout for the virtual network. This physical layout needed to be created before the virtual network to be completed in OPNET. The last software package used for this project is SolarWinds. This software is a subnet calculator that was used to confirm the IP Addresses used for the logical layout of the project. The logical layout was planned and worked out before using this subnetting calculator to confirm that the addresses were correct. 16

4.4 Hardware The main piece of hardware used for this project was a computer that was already purchased to complete all the testing in OPNET. This computer had all of the software loaded onto it that was necessary to complete this project. The only other hardware for this project was a 60 gigabyte external hard drive. This hard drive was used to store all of the data saved from OPNET. 5. Proof of Design The next section shows how the deliverables for this project were completed, and the steps taken to completion. 5.1 Scenarios There are two main scenarios that were used for the testing on this virtual network. The main scenarios are the Campus Backbone and the Edge Distribution. View Figures 5 and 6 to see the Campus Backbone and Edge Distribution. Figure 5 shows the Campus Backbone with the building distributions. Figure 6 displays the Edge Distribution with the four DMZs. 17

Figure 5 Campus Backbone Figure 6 Edge Distribution The Scenarios are not limited to just the two that appear in the screenshots. Certain failures were applied to these two scenarios to see how the routing protocols 18

affected the network traffic. View Figures 7 and 8 to see the failures that were applied to the Campus Backbone and Edge Distribution. In the Campus Backbone scenario, one of the Campus Backbone switches was failed, and one of each of the sets of building distribution routers was failed. The failure is makes the devices act as if they were powered off, or had a hardware failure. For the Edge Distribution, one of the edge distribution switches was failed. Figure 7 Campus Backbone with Failure 19

Figure 8 Edge Distribution with Failure 5.2 The Complete Virtual Network The complete virtual network was a necessary deliverable to complete the other objectives in this project. Figure 9 displays the complete virtual network based on Cisco s Enterprise Composite Network Model. This model consists of the campus backbone, building distributions, edge distribution, and the four DMZs. The four DMZs consist of a service provider edge, an e-commerce edge, an internet edge, and a remote access edge. The service provider edge contains a frame relay connection, ATM connection, and a Point-to-Point Protocol (PPP) connection. The e-commerce edge provides a means of revenue for a company. The internet edge provides the users on the network with a connection to the outside world. The remote access edge provides remote connections for users at home or away from the office. 20

Figure 9 Complete Virtual Network 5.3 Convergence Times Convergence time is known as the time it takes for all the routers and switches in the network to know exactly what connections are active on the network. Another good way to explain convergence time is that it is the time it takes for all the routers and switches on the network to know the current network topology. Convergence time is something that exists with the EIGRP routing protocol and not RIP. Convergence time is possible because of the Diffusing Update Algorithm (DUAL). This algorithm provides EIGRP with the ability to calculate the lowest cost path from a device if it fails. Convergence happens before a device even fails, so that EIGRP can adjust to a failure right away. It is interesting to see how fast different sections of the network converge when set up with different parameters. Not only is it interesting, but it also provides valuable data for supporting EIGRP as a superior routing protocol (10 pgs. 235-250). 21

See Figure 10 to see how the convergence time looks on the Campus backbone with general traffic running. This convergence time is about.0038 of a second. Figure 10 Covergence time Campus Backbone General Traffic Heavy Load The convergence time with the Campus Backbone running general traffic on light load is just about the same as it is with a heavy load. There was a noticeable difference when running heavy traffic on the campus backbone with failure. See Figure 11 to view the convergence time on the Campus Backbone with Failure. This convergence time is slightly quicker with no failure because EIGRP is converging with fewer devices. Also, EIGRP already knows the shortest paths to take when there is a failure, so the convergence time is actually quicker with failure. This convergence time does not change much when changed to a light load. The convergence time in this scenario is about.00225 of a second. 22

Figure 11 Convergence time - Campus Backbone with Failure General Traffic Heavy Load The Edge Distribution convergence time with general traffic on heavy load is much slower to converge. This convergence time is about.175 of a second. View Figure 12 to see the convergence time on the Edge Distribution with general traffic on heavy load. This convergence does not noticeably change with a light load. 23

Figure 12 Convergence time Edge Distribution General Traffic - Heavy Load The convergence time on the Edge Distribution with failure running general traffic with a heavy load was slightly faster than without the failure. This is consistent with the results from the Campus Backbone. The convergence time is about the same with a light load running on this scenario, which is also consistent with the Campus Backbone. See Figure 13 for the convergence time on the Edge Distribution with Failure running general traffic with a heavy load. 24

Figure 13 Convergence time - Edge Distribution with Failure General Traffic Heavy Load 5.4 Logical Layout If this network is going to exist theoretically, then it was important for it to exist logically. See Figure 14 for the IP Address layout for this virtual network. This IP layout was planned out, and then confirmed with the SolarWinds subnetting calculator. The logical layout only serves the purpose to show that this network could exist physically if the devices were available. The devices were assigned IP Addresses from the available range of valid IP Addresses. There were eight different subnets based on the design of the network. Those subnet address numbers can be viewed in Figure 14. 25

Host Address Email Server 170.1.16.1 DNS Server 170.1.16.2 File & Print Server 170.1.16.3 Directory 170.1.16.4 Database 170.1.16.5 Legacy 170.1.16.6 Server Farm Switch 1 170.1.16.7 Server Farm Switch 2 170.1.16.8 Campus Backbone Switch 1 170.1.16.9 Campus Backbone Switch 2 170.1.16.10 Building Distribution Router 1 (A) 170.1.32.1 Building Distribution Router 2 (A) 170.1.32.2 Workstation Switch 1 (A) 170.1.32.3 Workstation Switch 2 (A) 170.1.32.4 Workstation Switch 3 (A) 170.1.32.5 Workstation Switch 4 (A) 170.1.32.6 LAN 1 (A) 50 PCs 170.1.32.7-170.1.32.56 LAN 2 (A) 50 PCs 170.1.32.57-170.1.32.106 LAN 3 (A) 50 PCs 170.1.32.107-170.1.32.156 LAN 4 (A) 50 PCs 170.1.32.157-170.1.32.206 Building Distribution Router 1 (B) 170.1.48.1 Building Distribution Router 2 (B) 170.1.48.2 Workstation Switch 1 (B) 170.1.48.3 Workstation Switch 2 (B) 170.1.48.4 Workstation Switch 3 (B) 170.1.48.5 Workstation Switch 4 (B) 170.1.48.6 LAN 1 (B) 50 PCs 170.1.48.7-170.1.48.56 LAN 2 (B) 50 PCs 170.1.48.57-170.1.48.106 LAN 3 (B) 50 PCs 170.1.48.107-170.1.48.156 LAN 4 (B) 50 PCs 170.1.48.157-170.1.48.206 Building Distribution Router 1 C 170.1.64.1 Building Distribution Router 2 C 170.1.64.2 Workstation Switch 1 C 170.1.64.3 Workstation Switch 2 C 170.1.64.4 LAN 1 C 100 PCs 170.1.64.5-170.1.64.104 LAN 2 C 100 PCs 170.1.64.105-170.1.64.204 Corporate Router 170.1.80.1 Edge Distribution Switch 1 170.1.16.11 Edge Distribution Switch 2 170.1.16.12 E-Commerce Internal Router 170.1.96.1 E-Commerce Internal Firewall 170.1.96.2 E-Commerce Local Switch 170.1.96.3 E-Commerce Web Server 170.1.96.4 E-Commerce DMZ Firewall 170.1.96.5 E-Commerce Application Server 170.1.96.6 E-Commerce Internet Router 170.1.96.7 E-Commerce Database 170.1.96.8 Internet Internal Router 170.1.112.1 Internet Internal Firewall 170.1.112.2 Internet Local Switch 170.1.112.3 Internet Cache Server 170.1.112.4 Internet DMZ Firewall 170.1.112.5 Internet Public Server 170.1.112.6 Internet - Internet Router 170.1.112.7 Remote Access Internal Router 170.1.128.1 Remote Access Firewall 170.1.128.2 Remote Access Local Switch 170.1.128.3 Remote Access IDS 170.1.128.4 Remote Access VPN 170.1.128.5 Subnet Number Subnet Broadcast Address Range of Valid Addresses 170.1.16.0 170.1.31.255 170.1.16.1-170.1.31.255 170.1.32.0 170.1.47.255 170.1.32.1-170.1.47.255 170.1.48.0 170.1.63.255 170.1.48.1-170.1.63.255 170.1.64.0 170.1.79.255 170.1.64.1-170.1.79.255 170.1.80.0 170.1.95.255 170.1.80.1-170.1.95.255 170.1.96.0 170.1.111.255 170.1.96.1-170.1.111.255 170.1.112.0 170.1.127.255 170.1.112.1-170.1.127.255 170.1.128.0 170.1.143.255 170.1.128.1-170.1.143.255 Figure 14 IP Address Layouts 26

5.5 Application Delay Times The applications delay times on this virtual network are consistent with being quicker with EIGRP. This is very noticeable with heavy loads on the network. With lighter loads the application response time is not that noticeably quicker with EIGRP, but it is still often quicker. View Figure 15 to see how the Database Entry response time with a heavy load going over the Campus Backbone. This Database Entry response time was for a simulated time of six minutes, and it is noticeably quicker with EIGRP. Figure 15 Database Entry response time Campus Backbone Heavy Load The FTP traffic is also noticeably quicker with EIGRP when a heavy load is placed on the Campus Backbone. See Figure 16 to see the difference in the FTP traffic on this network. This screenshot shows the FTP download response time, and the upload response time is about the same. 27

Figure 16 FTP Download response time Campus Backbone Heavy Load With a light load running on the Campus Backbone, the results were slightly different than Figure 15. The Database Entry response time is still quicker with EIGRP, but not as noticeably. See Figure 17 to view the differences between EIGRP and RIP with Database Entry on light load. 28

Figure 17 Database Entry response time Campus Backbone Light Load When looking at the Edge Distribution scenario, similar results were obtained. With general traffic running over the network at a heavy load, there is a noticeable difference between the Database Entry response time with EIGRP and RIP. EIGRP gets a better response time that RIP with a heavy load running on the Edge Distribution. See Figure 18 to view the differences between EIGRP and RIP with the Database Entry response time. 29

Figure 18 Database Entry response time Edge Distribution Heavy Load With a light load flowing over the network, the differences in the Database Entry response time is not as noticeable when compared to Figure 18. This is also consistent with the results from the Campus Backbone. View Figure 19 to see the Database Entry response time with a light load on the Edge Distribution. 30

Figure 19 Database Entry response time Edge Distribution Light Load 5.6 Overall bit flow The overall bit flow shows very large differences between RIP and EIGRP. The differences between EIGRP and RIP with the overall bit flow was outstanding. In every scenario with different types of traffic flowing over the network, RIP always sends at least twice as much traffic over the network. The main reason for this traffic difference is because RIP sends automatic updates every thirty seconds. These automatic updates are large packets that include the routing information from a router or switch. View Figures 20 and 21 to see the differences between the overall bit flow on the network with the Campus Backbone on a heavy load. 31

Figure 20 Overall Bit Flow Campus Backbone - General Traffic Heavy Load Figure 20 shows the overall bits per second on the Campus Backbone. RIP is exceeded 50,000 bits per second every thirty seconds, while EIGRP sends fewer bits per second at about 3,000 bit per second. The results are about the same when a light load is applied. The traffic sent by RIP is averaging over twice as much traffic. 32

Figure 21 Overall Bit Flow Campus Backbone - Video Conferencing - Traffic Only When testing one type of traffic out by itself, the results are consistent to when general traffic runs on the Campus Backbone. Figure 21 shows how much traffic is flowing over the Campus Backbone with video conferencing traffic. RIP is still spiking over 50,000 bits per second, while EIGRP continues to flow at about 3,000 bits per second. The Voice over IP (VOIP) traffic was consistent with the video conferencing results. The Edge Distribution had results that are consistent with the Campus Backbone. The only major difference is that RIP was sending even more traffic on the Edge Distribution than EIGRP. See Figure 22 to see how drastically the difference is between the traffic flowing on the Edge Distribution with a heavy load. RIP reached over 75,000 bits per second, while EIGRP stayed at about 3,000 bits per second. The Video 33

Conferencing and VOIP traffic gave the same results when they were running individually on the network alone. Figure 22 Overall Bit Flow Edge Distribution General Traffic Heavy Load 5.7 Results with a Failed Node Another important deliverable for this project is to see how this network reacts with a failed node when running RIP and EIGRP. Different failures were set up for these scenarios. On the Campus Backbone, one of the Campus Backbone Switches was failed, and one of each of the pairs of building distribution routers was failed. It is important to view the results with a failed node because a network engineer would like to know how a network will respond to failure. How long will the network be down, or unresponsive? These results help clarify how a network will respond with certain failures. See Figure 23 34

to view the results with a failure on the Campus Backbone with Database Entry traffic sent. Figure 23 Results with a Failed Node Database Entry traffic sent Campus Backbone Heavy Load It is clear to see that RIP sends more Database Entry traffic than EIGRP with a failure on the Campus Backbone. Figure 23 displays the results of Database Entry traffic running over the Campus Backbone with several failures. The results with Database Query traffic are about the same. With a light load on this scenario the results are also very similar, but the difference is not as noticeable. Since the difference is not as noticeable with a light load, EIGRP shows to be more superior when heavy loads are applied to the network. The Edge Distribution with a failed node is consistent with the results from the Campus Backbone. With one of the Edge Distribution switches failed out, the Database 35

traffic is higher with RIP. View Figure 24 to see the difference between the Database Entry traffic sent over the Edge Distribution with a failed node. Figure 24 Edge Distribution with Failure - Database Entry traffic sent - General traffic - Heavy Load 5.8 VOIP Traffic The last important deliverable is the Voice over IP (VOIP) traffic. This traffic effects how phone calls are sent and received throughout a network. If there is delay with this traffic, then a phone call will receive latency, or the call will be dropped. It is important for upper management to consider how vital their phone calls are. Is this a business transaction that could possible be dropped? If so, then the following results for this network should be considered. See Figure 25 to view how VOIP packet delay variation is different between RIP and EIGRP. 36

Figure 25 VOIP packet delay variation Campus Backbone From Figure 25, it is clear to see that there is a delay difference between RIP and EIGRP. View Figure 26 to see the packet end-to-end delay with VOIP running over the Campus Backbone. 37

Figure 26 VOIP packet end-to-end delay Campus Backbone The Edge Distribution is consistent with the Campus Backbone when it comes to VOIP traffic. RIP is still showing to be slower than EIGRP when dealing with VOIP. View Figure 27 to see the differences between RIP and EIGRP with VOIP running on the Edge Distribution. 38

Figure 27 VOIP Packet end-to-end delay Edge Distribution When running tests to see how VOIP traffic reacts with a failure on the Edge Distribution, it is another consistent result with the Campus Backbone. RIP shows a slower response than EIGRP when one of the Edge Distribution switches is failed. See Figure 28 and 29 to view the differences between RIP and EIGRP with VOIP traffic running over the Edge Distribution with failure. Figure 28 show a small difference between RIP and EIGRP, but Figure 29 show RIP displaying packet delay while EIGRP almost has no delay. 39

Figure 28 VOIP Packet Delay Variation Edge Distribution with failure Heavy Load Figure 29 VOIP Packet Delay Variation Edge Distribution with failure - Light Load 40

One last thing to consider about VOIP traffic is that the overall traffic being sent with only VOIP traffic running on the network. This traffic difference between RIP and EIGRP is staggering. View Figure 30 to see the differences in traffic running on the Edge Distribution with VOIP traffic only. This screenshot shows RIP sending over 65,000 bits per second every thirty seconds. Meanwhile, EIGRP averaged about 3,000 bits per second. This was an important fact because it shows how inconsistent RIP can be. With VOIP traffic it is best to have a consistent routing protocol so that no latency or lag will be present on phone calls. Figure 30 VOIP traffic only Edge Distribution 6. Conclusions and Recommendations The following sections will discuss the final conclusions and recommendations that explain why this research is important and useful. 41

6.1 Conclusions The advantages of using EIGRP on an Enterprise Network stand out in four major points. These points are the application delay, overall bit flow, results with a failed node, and VOIP traffic. The tests run on this network conclude that the application response time is quicker with EIGRP. The difference in the overall bit flow is outstanding. RIP sends at least twice as much traffic over the network when compared with EIGRP. When a router or switch fails on the network, EIGRP responds much quicker than RIP. The network can respond to failure right away with EIGRP, while RIP may take several minutes to adjust to a failure. Finally, the benefits when using EIGRP with VOIP should give upper management comfort because EIGRP has less delay, which means less latency and fewer dropped calls. 6.2 Recommendations The information presented in this study can be used to help network engineers choose what routing protocol to use over a network. There are many advantages to EIGRP, but if a small network is being implemented, then RIP can still be a good choice. RIP has been known to be very successful on small networks because it is easy to set up, and it is easy to troubleshoot. RIP is also stable under a light load. When a network grows to what is considered mid-to-large in size with a high traffic flow, it is important to consider another routing protocol such as EIGRP. EIGRP can handle thousands of routing nodes, and it is meant to be used for very large networks. 42

References 1. Answers.com. 2006. Answers Corporation. October 30, 2006. http://www.answers.com/topic/enhanced-interior-gateway-routing-protocol 2. Black, Uyless. IP Routing Protocols. Upper Saddle River, NJ: Prentice-Hall Incorporated, 2000. http://proquest.safaribooksonline.com/0130142484 3. Cisco. 2006. Cisco Systems Incorporated. October 12, 2006. http://www.cisco.com/univercd/cc/td/doc/cisintwk/ito_doc/rip.htm http://www.cisco.com/univercd/cc/td/doc/cisintwk/ito_doc/en_igrp.htm 4. Data Network Resource. 2006. Rhys Haden Incorporated. November 10, 2006. http://www.rhyshaden.com/eigrp.htm 5. Doyle, Jeff. CCIE Professional Development Routing TCP/IP, Volume I, Second edition. Indianapolis, IN: Cisco Press, 2006. http://proquest.safaribooksonline.com/1587052024 6. Holden, Greg. Guide to Firewalls and Network Security. Canada: Thomson Learning Incorporated, 2004. 7. Malhotra, Ravi. IP Routing. Sebastopol, CA: O Reilly & Associates Incorporated, 2002. 8. Microsoft. 2006. Microsoft Corporation. October 30, 2006. http://www.microsoft.com/technet/prodtechnol/windows2000serv/reskit/intwork/i nae_ips_nxtl.mspx?mfr=true 9. OPNET. 2007. OPNET Technologies, Inc. May 18, 2007. http://www.opnet.com/ 43

10. Oppenheimer, Priscilla. Top-Down Network Design. Indianapolis, IN: Cisco Press, 2004. 11. SolarWinds. 2006. SolarWinds Network Management Solutions. November 14, 2006. http://www.solarwinds.net/ 12. Sportac, Mark. IP Addressing Fundamentals. Indianapolis, IN: Cisco Press, 2003. http://proquest.safaribooksonline.com/1587050676 13. Stewart, Brent. CCNP ONT Quick Reference Sheets. Indianapolis, IN: Cisco Press, 2006. 14. TechRepublic. CNET Networks Incorporated. October 30, 2006. http://articles.techrepublic.com.com/5100-1035-6089187.html 15. Zinin, Alex. Cisco IP Routing: Packet Forwarding and Intra-domain Routing Protocols. United States: Addison-Wesley Incorporated, 2002. 44