1 Andy Davis Akamai Technologies 1400 Fashion Island Blvd San Mateo, CA EdgeComputing: Extending Enterprise Applications to the Edge of the Internet Jay Parikh Akamai Technologies 1400 Fashion Island Blvd San Mateo, CA William E. Weihl Akamai Technologies 1400 Fashion Island Blvd San Mateo, CA ABSTRACT Content delivery networks have evolved beyond traditional distributed caching. With services such as Akamai's EdgeComputing it is now possible to deploy and run enterprise business Web applications on a globally distributed computing platform, to provide subsecond response time to end users anywhere in the world. Additionally, this distributed application platform provides high levels of fault-tolerance and scalability on-demand to meet virtually any need. Application resources can be provisioned dynamically in seconds to respond automatically to changes in load on a given application. In some cases, an application can be deployed completely on the global platform without any central enterprise infrastructure. Other applications can require centralizing core business logic and transactional databases at the enterprise data center while the presentation layer and some business logic and database functionality move onto the edge platform. Implementing a distributed application service on the Internet s edge requires overcoming numerous challenges, including sandboxing for security, distributed load-balancing and resource management, accounting and billing, deployment, testing, debugging, and monitoring. Our current implementation of Akamai EdgeComputing supports application programming platforms such as Java 2 Enterprise Edition (J2EE) and Microsoft s.net Framework, in large part because they make it easier to address some of these challenges. In the near future we will also support environments for other application languages such as C, PHP, and Perl. Categories and Subject Descriptors C.2.4 [Computer-Communication Networks]: Distributed Systems distributed applications, distributed databases, Web services; D.2.11 [Software Engineering]: Software Architectures Java, languages,.net, patterns; D.2.5 [Software Engineering]: Testing and Debugging debugging aids, diagnostics, distributed debugging. General Terms Management, Performance, Design, Reliability, Security, Copyright is held by the author/owner(s). WWW 2004, May 17 22, 2004, New York, New York, USA. ACM /04/0005. Standardization, Languages Keywords Edge computing, grid computing, distributed applications, splittier applications, web applications, N-tier applications, utility computing, Internet applications, Web services 1. Introduction The Web has evolved dramatically from its beginnings as a global publication mechanism. Businesses and other organizations are increasingly using the Web for online business processes, ranging from supply chain management to product configurators to customer and partner portals. The key technological shift underlying this is an increasing reliance on interactive Web applications in a strategy to grow business and/or increase operational efficiency by reaching more customers and partners. These applications serve many of the same functions as older green-screen applications or clientserver applications, but users now interact with them via the Web. The Internet, and by extension, the Web, is a far from optimal delivery vehicle for interactive applications. Businesses use the Internet because it is better than doing nothing: they can reach more users more cheaply than with alternatives such as private networks, dial-up, phone, or fax. But many applications delivered over the Web still suffer from performance, reliability, and scalability problems. These problems are not new; they motivated the first wave of Content Delivery Networks (CDNs) in the late 1990 s. However, the problems may be more severe now than in the early days of the Web, because the business processes implemented by many of the Web-facing applications in use today are often more missioncritical to businesses. CDNs solved the problems of the Web for static content by delivering content from caches at the edge of the Internet, close to end users. The large scale and distributed nature of CDNs permits greater reliability. Also, delivering content from edge servers ensures faster download times, particularly if those servers are carefully chosen based on network conditions and other factors. Further, the ability to allocate resources dynamically and automatically permits enormous scalability in response to marketing successes, news events, and other events leading to flash crowds. As web sites have moved to greater use of interactive applications, CDNs have continued to evolve. Caching static
2 content provides little help for an interactive application. The embedded images and other objects may be delivered from an edge cache, but if the base HTML page is generated by an application, user requests must still travel a long distance on the Internet to contact the origin application that generates the content. Congestion and outages on networks cause some users to see download times for a page in the 10 s to 100 s of seconds enough to make a site virtually unusable for those users. Additionally, capacity limitations on the enterprise data center infrastructure limit the number of users that an application can support; for consumer-facing applications the load is particularly hard to predict, resulting in money spent on excess capacity that remains idle much of the time but still insufficient to support peak loads. We have developed a distributed application service, Akamai EdgeComputing, to address these problems. EdgeComputing brings to applications the same advantages that CDNs provide for static content. Akamai s EdgeComputing distributed application service is a form of utility or grid computing (see the Global Grid Forum  and the Globus alliance  for information on many current grid computing efforts). Unlike most grid computing systems, however, it is focused on interactive business applications. It is also globally distributed to ensure that applications run close to their end users, automatically providing capacity both when and where it is needed. Using EdgeComputing, parts of a Web application and in some cases the entire application can be distributed across the Akamai network. Many client requests can be processed completely at the edge, avoiding wide area network communication altogether. For requests that require communication with the enterprise data center (e.g., to talk to a transactional database, or to interact with legacy systems), only the raw data needs to be exchanged, not the entire HTML page, reducing the amount of data sent long-haul by one to two orders of magnitude. In addition, the same network of servers can be used to optimize the path taken over the Internet between the edge and the enterprise, avoiding congestion and outages in realtime and further reducing response time. At its core, our EdgeComputing service requires the ability to virtualize server resources. There are many ways to do this, including virtual machine monitors (e.g., VMware , Xen ), user-mode Linux , and binary code rewriting (e.g., Etch , ATOM/OM ). For performance and manageability reasons, we chose not to provide each customer with a set of virtual machines for its applications. Instead, we run each customer s applications in separate application server processes, relying on kernel and operating system functionality to isolate customers from each other and to prevent runaway use of server resources. We then provide a customer application console through which a customer can view a summary of all of his running application instances, and when necessary drill down to a detailed look at individual instances. Initially, we have implemented EdgeComputing for J2EE and.net, which provide standard programming environments that simplify the deployment and management of clustered Web applications. Our current implementation supports Apache s Tomcat server and IBM s WebSphere Application Server; support for Microsoft s.net framework will be available later in 2004, followed by support for native application language environments (C, Perl, etc.). Our original motivation for a distributed application service was to solve the performance, reliability, and scalability problems plaguing Web applications. As we discuss below, EdgeComputing has enabled our customers to ensure truly interactive sub-second response time to all users, with far greater reliability and scalability than they could cost-effectively achieve on their own. In addition, EdgeComputing has resulted in unexpected benefits, though viewed in retrospect the benefits are not so surprising. Time to market has been greatly reduced for many applications because customers no longer have to deal with capacity planning, acquiring, and provisioning infrastructure, thus eliminating significant risk, cost, headaches, and delays. If an enterprise still maintains a central infrastructure, they need less. Customers can focus on what they want the application to do and on building it. Deployment then involves simply publishing the application to Akamai, which manages the application infrastructure within an on-demand environment. Customers also can worry less about optimizing applications for scale, although they (and we) still need to worry about performance issues overall. 2. The EdgeComputing Model Our EdgeComputing distributed application service is not a new programming model or a new set of programming APIs; it is a new deployment model for Web applications. Today, most Web applications are deployed within a single data center, perhaps in a cluster of machines connected by a high-speed LAN. To use EdgeComputing, a site developer typically must split the application into two components: an edge component and an origin component. The code in the edge component is deployed onto Akamai s network of servers distributed around the world (more on how this is done below); the origin part is deployed in the traditional manner within the central data center. Some applications naturally and easily divide into edge and origin components; others require some redesign. We discuss this issue in more detail later in the paper. We have implemented a replication subsystem for session state, permitting an application s edge components to maintain peruser state that remains available even when users are mapped to different servers; more detail on session management is in Section 6. Edge components can also use our NetStorage facilities (a distributed collection of storage silos that provides automatic replication of data) to store information submitted by users for later processing by the origin components. Our management interfaces support staged deployment for applications. Typically, a site developer tests an application internally on test machines before deploying it to Akamai. After the application passes internal testing, the developer deploys it across the Akamai network in stages, watching for problems at each stage. More specifically, deployment involves several steps. (For concreteness, we focus here on J2EE applications.) 1. A site developer packages the edge part as a WAR (Web ARchive) file. 2. The developer publishes the WAR file to Akamai through the Akamai portal (via a Web page or Web service). 3. The Akamai portal scrubs the WAR file to ensure that it can run on our edge application servers. Akamai validates the WAR file and makes sure the components
3 are compiled for the correct application server platform. 4. The developer uses the Akamai portal to control the deployment of the application, first deploying it to a test network of Akamai machines that are identical to our production machines except that they do not serve live traffic, then to a few machines serving live traffic, then to a small but significant percentage of the network, and finally to the entire network. At each stage of the process, the developer can drive test load against the application and watch for errors, alerts, and other indications that there are problems with the application. 5. If problems are found, the developer can abort the deployment and revert back to the older application. Deployment occurs in stages such that multiple versions of an application may be live simultaneously, a process that limits the impact of errors to as few users as possible. Applications belonging to different customers may run simultaneously on the same machine. We have implemented a security sandbox to ensure that applications cannot interfere with one other or with the underlying operation of our servers. Security sandboxes are discussed in Section 4.1. Server resources on the Akamai network are managed automatically. When the load on an application increases, additional instances of the application are started on additional servers near the users making the requests, and requests are directed to those servers. When the load drops, application instances may be stopped automatically. This automatic resource management and load balancing system is discussed in Section 4.2. Logically, a site developer can view the Akamai system as a large virtual cluster, distributed around the world. Server resources can be added or removed from the distributed cluster at any time, dynamically and automatically. To enable debugging in such a system, our portal provides a comprehensive view of all running instances of an application. Debugging is discussed in more detail in Section 4.3. Finally, billing for our EdgeComputing service is based on usage, measured in requests per month. Applications are allocated a certain amount of bandwidth, memory, CPU, and other resources. Applications can be given more than the standard allocation, typically for an additional application resource fee. An alternative would be to allow unlimited resource usage, and to bill for what is used. However, most customers have indicated a preference for a model in which usage per request is limited in order to avoid receiving an enormous bill at the end of the month because of an infinite loop or some other bug that creates runaway usage. 3. Example Edge Applications We launched EdgeComputing as a service in early As of this writing, we have a large number of customers running a wide range of applications on the service. Those applications fall into several categories, summarized in the following subsections. 3.1 Content Aggregation The simplest category of applications involves applications that have no internal databases, but simply aggregate and format content from other sources. This category includes portals, which aggregate news, search, and other data sources to present a unified interface. This content can be transformed depending on the end user device. These applications are simple to run on the edge; data can be retrieved from the various sources using HTTP or Web services, formatted on the edge, and then returned to the user. If the data from the various sources is cacheable, most user requests can be processed completely on the edge without any long-haul communication. Note that some portals today permit customization based on a user profile kept in a database. Today, that database would be maintained in our customer s data center. User profiles can be retrieved from the database over the Internet (typically through a proxy servlet or similar mechanism running at the origin in front of the database), and then used to customize the page returned to the user. With some care, user profiles can also be cached on the edge, allowing most communication with the origin to be avoided altogether. Another application in this category found on many sites is a store locator application, which allows a user to enter a location, for example, by city or zip code, and find out what stores are near that location. In many cases, the real work of this application is done by a Web service such as MapPoint ; the application on the customer site simply formats the results returned by the service. This is really a special case of portals, requiring only a single data source. 3.2 Static Databases Applications with static databases can often be run entirely at the edge. Applications today are limited to databases of moderate size (100 s of megabytes). Examples include: - Store locator the part often done by MapPoint, where the database involves the actual stores and their locations as well as the geographical information needed to determine proximity. - Product catalogs except for real-time inventory information, the information in most product catalogs is relatively static and can easily be cached on the edge servers. For J2EE applications, we are working with IBM to deploy Cloudscape , a pure Java relational database, to enable these kinds of databases to be run alongside the edge application, accessed using JDBC. Note, that the Cloudscape edge database technology is applicable to many other types of edge applications, not just product catalogs. - Site search the index for searching a site can typically be cached on the edge. - Product configurators again, the database containing product information and the business rules controlling allowable configurations and associated pricing can often be cached on the edge. Configurators are a little different from the other applications in this category, since they often require client session state to keep track of the configuration selections already made by the user. We replicate session state across multiple servers in order to handle when a user gets mapped to different machines over time without losing the user s session.
4 3.3 Data Collection Some applications, such as registration (for products, college courses, etc.), online applications (where one applies for something, such as college or a credit card, not in the sense of computer applications ), and customer service (submitting a question via a web form) involve collecting data from forms filled in by users. If the application is a single-page form, the data can be collected on the edge and sent via HTTP POST (or a Web service call) directly to the origin, or the POST data could be sent to our NetStorage system. The NetStorage system replicates the data in files on disk for later retrieval by the customer. If the application form involves multiple pages, the information entered progressively can be kept in session state on the edge. The choice of which form to present next a choice perhaps based on the data entered so far can be made on the edge, and the final data can again be stored on NetStorage or sent to the origin. Data validation can also be done at each step to filter out nonsensical data. As another option, if the data is only needed for later offline batch processing, the edge application server can log some data for each request. This logged data is collected from all servers. Later, the collated log files are made available for processing by the origin. Some news and portal sites, among others, use voting or polling applications to allow their users to express their opinion or preferences on various issues. This data can be collected and buffered for short periods on the edge and then sent to the origin. This allows the origin to service a small number of medium-tolarge requests instead of a very large number of small requests. This reduction in turn enables the application to scale to larger loads without significant dedicated central infrastructure. The origin can then publish the aggregated data to the edge, where the application can refresh its view every few seconds to give users virtually real-time polling results. Loads on this kind of application can be particularly hard to predict and thus prohibitively expensive to provision dedicated infrastructure for peak loads. 3.4 Two-Way Data Exchange A slightly more complex version of the data collection application involves data flowing in both directions withthe behavior of the application depending on the aggregated data published in real-time by the origin. One example is online ad serving in which the ad to be served depends on the available ad inventory and which ads have been served so far. Another example is contests, in which the probability of winning may depend on the number of people who have already won. As with the voting or polling application, buffering and aggregating data in the network for brief periods can reduce the central infrastructure required by several orders of magnitude, enabling a site to handle much higher peak loads at lower expense. In the case of one customer, a contest application was designed and developed to run on Akamai EdgeComputing; it used virtually no central infrastructure, but still supported over 70 million requests during the 5-hour contest. Without EdgeComputing, a great deal of time would have been required to procure and provision the infrastructure, and it is likely that the load would have exceeded the capacity in any event, since the actual load exceeded the expected load by roughly a factor of Complex Applications The final category is the catchall of more complex applications, including full e-commerce engines, supply chain management, customer relationship management, online banking, etc. Many of these applications have pieces that fall into the other categories above (e.g., e-commerce engines include product search, store locators, and other pieces that can be moved to the edge), but they also include pieces that rely on transactional databases. In such cases, the application must be split, typically to put the presentation layer on the edge and leave the business logic and data access layers at the origin. Even though some central infrastructure is still required, there can still be significant advantages in performance because only raw data needs to be retrieved from the origin, rather than full HTML pages. More information on best practices for how to handle these types of applications is provided in Section Technical Challenges In this section we discuss a number of technical challenges faced in implementing the EdgeComputing distributed application service, including security, debugging, accounting and billing, load balancing, resource management, and monitoring, and session state replication. 4.1 Security Customers EdgeComputing applications run on machines within the Akamai network, and multiple applications from different customers may run simultaneously on a machine. For this reason, each EdgeComputing customer application is executed within a security sandbox that prevents EdgeComputing applications from accessing unauthorized server resources and data, and also keeps each application from over-utilizing granted server resources. Our primary concern in isolating applications from each other is to protect against buggy code, not against malicious users; we have a business relationship with our customers that makes the deployment of malicious code unlikely. The security sandbox performs access control checks each time an application tries to access a resource and checks limits on resource usage. For example, if an EdgeComputing application is allowed access to an area on disk, the sandbox prevents the application from reading or writing data outside the allotted disk area, limits the rate of disk access, and enforces a limit on the total amount of data stored on disk. The sandbox also limits the hosts to which the application can communicate, and it limits the number of network connections the application can have open at any given time. To provide access control and resource management, the security sandbox makes use of facilities from the operating system such as file system quotas, user/group id permissions, and custom sandbox modules. In addition to these facilities, the sandbox can be extended with finer granularity of access control and sandboxing by making use of security capabilities provided by the application server, such as the J2EE Security Manager and policies. To achieve better isolation, each customer s EdgeComputing application is run in a separate process on each machine. This isolates any ill effects from a misbehaving application and prevents that misbehaving application from hampering the performance of other applications running on the system. Before an EdgeComputing application is launched, libraries and
5 configuration files needed by the process are stored in an area of disk allocated to the process, and permissions and quotas are set up as necessary. Also, the application s process is configured such that it cannot access files or directories outside of the allocated area on disk. Once a process is launched for an application, the custom sandbox modules track system calls made by the server process. The sandbox modules are used to throttle access to resources and to report resource usage information to an external monitoring process that monitors all EdgeComputing applications running on the machine. The monitor process may restart, throttle, or terminate a process that tries to access forbidden resources or over-utilizes resources on the machine. It also monitors the time used by the process for each user request, and it may terminate or restart the process if requests take longer than a configurable threshold. Since edge applications can also use session data (managed and replicated by the Akamai EdgeComputing systems), the sandbox also enforces limits on session object size to prevent overconsumption of disk and memory for storing and replicating session data. Further, the security sandbox ensures that application session objects are accessible only from the appropriate customer application by encrypting the session data using separate customer specific keys. 4.2 Load Balancing, Resource Management, and Monitoring The problem of preventing overload of any particular server or application can be challenging enough within a single data center. In a globally distributed system, the problem is magnified. The Akamai system automatically monitors client traffic, network conditions, and application performance and automatically distributes traffic and applications as needed to prevent overload of any particular server or application instance. Logically, the Akamai network can be viewed as a single virtual multi-computer that can run multiple EdgeComputing processes. At a more detailed level, the Akamai network is made up of many individual Akamai edge servers. Each edge server is capable of starting, stopping, and monitoring the execution of an EdgeComputing process. Load balancing is done hierarchically, first among edge server groups (ESGs), and then within ESGs. Within an ESG, load balancing focuses on ensuring that each application is running on enough servers in the ESG to handle the load for that application directed at the ESG. Among ESGs, load balancing focuses on distributing load for different applications to ensure that no ESG runs out of capacity, at the same time optimizing for network conditions and end user performance. At the top-level, load is dynamically mapped among ESGs using DNS, much as in the original Akamai CDN. Within an ESG, load is redistributed both using DNS (by returning more or fewer IP addresses in response to DNS requests) and by tunneling requests, as needed, from one server to another to ensure a relatively even distribution of load among instances of an application. Within an ESG, the load-balancing algorithm uses detailed information about resource consumption by each EC application instance. Within the ESG, load-balancing agents aggregate load and capacity data and report the summarized data to the top-level load-balancing agents. In addition, the agents analyze the local ESG data to determine how many instances of each application should be running in the ESG and whether load should be shifted within the ESG. As mentioned previously, each edge server runs a monitor process that monitors all EdgeComputing applications running on the server. The monitor process gathers resource usage, health and event information from the OS as well as from Akamai components running in-process with each EdgeComputing application. This data is fed to the load-balancing system as well as to the Akamai reporting system. The reporting system aggregates data from all machines in the network, providing a real-time monitoring data view. The CPU, disk, and memory usage of each application is monitored and reported. Periodic checks are made to each EdgeComputing process to make sure that it is functioning properly. Some EdgeComputing processes can be deployed to provide a periodic heartbeat message to be sent to the monitor process as a proactive indication that it is alive. Events generated by the EdgeComputing process (perhaps in response to a message from the Akamai network) are also monitored and reported. For J2EE applications, a finer granularity of resource monitoring is possible. Because these application servers use Java blocking I/O APIs and a thread-per-request model, we can track and report the CPU used for each request. In addition, the memory free and total memory used inside the JVM is also monitored and reported. All of the aggregated monitoring data is reported from each EdgeComputing process running across the entire Akamai network. This data can trigger sophisticated alerts for the Akamai NOCC. The data related to the customer s application is also presented to the customer through an application console. Many types of alerts can be generated; examples include memory utilization inside a customer s JVM, EdgeComputing process restarts because of resource over-utilization, failed customer application installs, and failed application delivery. Our customer application console also contains real-time information for each running instance of the customer s applications, showing the current state of the process, memory and CPU usage, and the number of client requests served by the process. 4.3 Debugging Each EdgeComputing process can write logs to disk that can then be accessed through the customer application console/portal. This allows the EdgeComputing application developer to write diagnostic information such as stack traces and event messages to a log file that can be viewed through the Akamai portal. Customers can see data for each instance of their applications running on the Akamai network. They can drill down on a specific instance on a specific machine and retrieve log files from that machine. Many JVMs allow programmatic generation of diagnostic information about the JVM and the application running inside it. This information includes thread dumps and JVM heap dumps. Our EdgeComputing implementation for J2EE allows this information to be requested through the customer application console for a specific application instance; the JVM writes the information to a file that can then be accessed by the customer in
6 the same manner as standard debug log files. A customer could retrieve this information from an application instance to diagnose a problem flagged by alerts or by other information available through the portal. Akamai provides a staging test network that replicates the production environment, except that it does not service live load. This permits functional and load testing of applications to be done in a safe environment that is as close as possible to the production environment. 4.4 Application Session State End user session state is a common aspect of Web applications developed using J2EE or.net (and others). This application state is conventionally stored in-memory in the application server process that is handling the end user request. By default, most application servers will not replicate or persistently store this session data, so it is not made available to other application instances in a clustered environment. This introduces a single point of failure for users mapped to a particular application instance, since if the server goes down, the client session data would be lost. In clustered enterprise environments, one can make use of in-memory replication sub-systems or a database to replicate and store the created session state. This can provide a higher level of fault tolerance to the application. Akamai EdgeComputing dynamically maps end users to the appropriate application instance across the entire network. This introduces an interesting problem of session affinity, since we can move end-user traffic from server to server. Having one centralized database to provide persistent session storage would add unnecessary latency in servicing the request and might even render worthless the benefits of moving an application to the edge. EdgeComputing provides an edge session replication system that allows for client session objects to be replicated across servers in an ESG as well as across ESGs. This system provides the application server session interface with an in-memory cache to reduce latency. The locally cached session object is checked with the system replicated session object to validate that the session object is current, and if so, the object can be used by the EdgeComputing application (thus, avoiding the deserialization of the object data). If the end user is mapped to another server, the application server on that server checks its local cache. Remapping occurs when machines suffer hardware or software failures, when the network between a user and a machine becomes congested, and when load needs to be redistributed to prevent machines from becoming overloaded. However, remapping is relatively infrequent, so it is likely that consecutive requests in a user s session will contact the same server. In the common case, the user s session object will be found in the local cache. The first time the user has been mapped to this machine, the application server must retrieve the user s session object from the session replication system. If the application modifies the session object, the updated object is sent asynchronously to the edge session replication system. 5. Edge Application Best Practices The EdgeComputing development model remains the same as for centralized applications; it does not require the use of any proprietary APIs. The deployment model changes, not the programming model. If an existing application follows wellknown component programming best practices, adapting it for EdgeComputing is easier. For example, EdgeComputing for J2EE enables the execution of J2EE Web tier application components JSPs, servlets, tag libraries, and JavaBeans. In order to provide more context, this section focuses on describing the best practices specifically related to development for EdgeComputing for J2EE using Tomcat or IBM s WebSphere Application Server. These best practices have been established working jointly with IBM and described in detail in . In general, EdgeComputing applications are architected as two cooperating sub-applications: an edge-side application running on Akamai and a corresponding origin-side application. 5.1 Presentation Components on the Edge The most common EdgeComputing model is to deploy the presentation components of an application onto the Akamai edge servers and to cache access to origin data via the Java Web services client model. Typically, the Web application will be developed using a framework based on the Model-View- Controller (MVC) architecture. Jakarta s Struts framework , a common open-source framework for building Web applications based on the MVC pattern, is well suited for EdgeComputing. Struts provides its own Controller component via a servlet. For the View, Strutsbased applications often leverage JSPs to create the end user response. The Model component is commonly represented by JavaBeans. These Model JavaBeans may be self contained or represent façades for other components like JDBC or EJB. The View and Controller components of a Struts application are good candidates for distributing on EdgeComputing. These components run on the Akamai edge servers and can interact with Model components running at the origin (EJBs). Depending on the functionality of your application, the extent to which these applications can move onto EdgeComputing will vary. The edge View and Controller components are bundled, along with other Java classes, into a Web application archive (WAR) and deployed onto the EdgeComputing network. With EdgeComputing today, EJBs commonly remain running at the origin and are made accessible to the edge application via Web services interfaces. A session bean façade can be used to expose the necessary business logic as a Web service to the edge application. The edge application makes a Web service call back to the origin to invoke the appropriate logic through a session bean façade, perhaps made visible through a servlet running on the origin application servers. An edge application can and should cache the results of SOAP/HTTP(S) requests to minimize interactions with the origin. In addition to Web services as a communication channel, our EdgeComputing distributed application service supports other standard communication protocols, including the following: - HTTP(S) An edge application can make HTTP(S) requests to the origin to request content or data objects. HTTP responses from the origin should be cached on the edge when possible, so content or data objects can be persisted on the edge across user requests to further reduce the load on the origin. A developer could leverage Akamai s content caching system for content objects (HTML fragments). Akamai provides an
7 invalidation control mechanism (via a Web service) to customers, so that customer can control the invalidation of any content objects cached on the Akamai edge servers. Additionally, an edge application can leverage caching Java objects in the edge JVM instance. - JDBC Akamai provides a JDBC driver that allows edge applications to tunnel JDBC queries to the origin via HTTP(S). If an application has already been developed using JDBC for data transaction with a database, the JDBC/HTTP mechanism will make it easier to adapt an application for the edge. Further, JDBC query responses can be configured to be cached on the Akamai edge servers similar to HTTP responses. - RMI Edge application components can use RMI tunneled over HTTP(S) to communicate with the origin component. In this configuration, a servlet runs at the origin, intercepts the RMI request from the edge, and translates the request into the appropriate method calls to the business tier. Rather than using RMI, JDBC, and other protocols in their native mode to communicate from the edge to the origin, we tunnel them over HTTP(S) to avoid problems with firewall configurations and to avoid requiring that enterprises make their databases and legacy systems directly accessible on the Internet. 5.2 Data Access on the Edge Any of the HTTP-based protocols above are useful interfaces that allow you to bridge your applications from the edge to the origin, but it is still important to avoid excessive communication for edge-origin requests, as otherwise end-user latency and origin utilization will increase. Because there is an absolute cost for every roundtrip from the edge to the origin, calls should be as coarse-grained as possible. Multiple requests to the origin that are required to service a single end-user request should be bundled, if possible, and edge caching should be used to store data and other objects across requests. The Akamai EdgeComputing platform provides caching mechanisms to persist application data to minimize the interaction and load on the origin infrastructure on a request-byrequest basis. - Client Session EdgeComputing supports replicated client sessions. Session objects are most commonly used to keep per-user state information. Akamai enforces a size limit to the actual serialized object data for performance reasons. - Client Cookies An edge application can store some user specific data in user cookies or URL tokens. Privacy and security concerns, however, may prohibit an enterprise from using this mechanism. - Object Caching As previously mentioned, an edge application can make requests for data or content (using HTTP, Web Services, or JDBC) and these objects can be stored in the edge node s cache. Another powerful edge data capability employs IBM s 100% pure Java Cloudscape DBMS to host infrequently changing, readonly data in an edge database. In this model, application data is exported as part of the application WAR file. By installing Cloudscape on the edge, even the Model components of an MVC application can be run on EdgeComputing. An edge application can make use of JavaBeans and JDBC as Model components with Cloudscape as the DBMS to further reduce the communication to the backend enterprise systems. One current limitation of this model is that any time the application data changes, the application must be redeployed to the EdgeComputing network. 6. Future Work At its core, EdgeComputing involves virtualizing system resources and allocating them dynamically to different applications. Simple virtualization, however, is not enough; for more than a few customers to be willing to use EdgeComputing, it must be easy to manage applications running on it. Our goal has been to make it as easy to manage an application running on EdgeComputing as it is to manage an application running on a cluster of machines in a central data center. This includes support for testing, monitoring, and debugging, as well as support for splitting an application into edge and origin pieces and orchestrating communication between the edge and the origin. We are currently working on a number of areas to simplify the management of distributed applications on our EdgeComputing service, including: - Further simplifying management of multiple versions of applications. - Providing deeper insight into the performance of applications on our edge servers to make it easier for customers to tune the performance of their applications. - Handling multiple versions of application servers. - Supporting the full J2EE specification, including EJBs and JMS. - Synchronizing session state between the edge component and the origin component of an application. - Making data caching and replication on the edge transparent for more patterns of data access. - Providing pre-integrated access to key services required by business applications (e.g., identity management, access control, and payment processing). 7. Summary and Conclusions Akamai s EdgeComputing service allows enterprises to deploy and run Web applications on a globally distributed computing platform. Running applications on servers at the edge of the Internet in our distributed application service provides many advantages for an enterprise including capacity on demand, better end user performance, higher application availability, and faster time-to-market. Development of this distributed application platform presented several new and interesting challenges. For example, customers need visibility into the behavior of their running applications, even though the machines that run a given application might change from day to day or even minute to minute. Similarly, applications must be sandboxed to ensure that one customer s application does not interfere with that of another customer.
8 There are several types of applications that can benefit from this distributed application platform. By following some basic edge application best practices, such as minimizing communication with the origin and caching data on the edge, many important business applications can leverage Akamai s EdgeComputing distributed application service. 8. ACKNOWLEDGMENTS Akamai s EdgeComputing service would not exist without the incredible efforts of the many people who have contributed to it, including Andy Berkheimer, Annie Boyer, James Chalfant, Eddy Chan, Andy Ellis, Liz Greene, Dan Hang, Marty Kagan, Nate Kushman, Erik Nygren, Nicole Peill, Vasan Pichai, M. C. Ramesh, Eddie Ruvinsky, Sudesh Saoji, Alex Sherman, Danner Stodolsky, Ashis Tarafdar, and Kieran Taylor. Our partners at IBM have also contributed enormously to the service by helping integrate IBM s WebSphere Application Server into our EdgeComputing service and by working with us on numerous sample applications and customer engagements. In addition, we are grateful to Gerry Cuomo, Michael Jacob, and Raymie Stata for helpful comments on this paper. Finally, we must acknowledge the early contributions of the late Danny Lewin, who believed from the beginning that EdgeComputing would be valuable to our customers. 9. REFERENCES  Barham, P., Dragovic, B., Fraser, K., Hand, S., Harris, T., Ho, A., Neugebauer, R., Pratt, I., and Warfield, A. Xen and the Art of Virtualization. ACM Symposium on Operating Systems Principles, Bolton Landing, NY, October 19-22,  Cuomo, G., Martin, B., Smith, K., Ims, S., Rehn, H., Haberkorn, M., and Parikh, J. WebSphere Capacity - On Demand: Developing EdgeComputing Applications. October (http://www7b.software.ibm.com/wsdd/library/ techarticles/0310_haberkorn/haberkorn.html)  Global Grid Forum.  The Globus Alliance.  IBM Cloudscape ibm.com/software/data/cloudscape/.  Jakarta Struts.  MapPoint service.  Romer, T., Voelker, G., Lee, D., Wolman, A., Wong, W., Levy, H., and Bershad, B. Instrumentation and Optimization of Win32/Intel Executables Using Etch Usenix NT Conference.  User-Mode Linux.  VMware.  Wall, D. W. Systems for late code modification. In Robert Giegerich and Susan L. Graham, eds, Code Generation - Concepts, Tools, Techniques, pp , Springer-Verlag, Also available as HP/Compaq WRL Research Report 92/3, May 1992.
Andy Davis Akamai Technologies 1400 Fashion Island Blvd San Mateo, CA 94404 EdgeComputing: Extending Enterprise Applications to the Edge of the Internet Jay Parikh Akamai Technologies 1400 Fashion Island
Oracle WebLogic Foundation of Oracle Fusion Middleware Lawrence Manickam Toyork Systems Inc www.toyork.com http://ca.linkedin.com/in/lawrence143 History of WebLogic WebLogic Inc started in 1995 was a company
SiteCelerate white paper Arahe Solutions SITECELERATE OVERVIEW As enterprises increases their investment in Web applications, Portal and websites and as usage of these applications increase, performance
Tuning WebSphere Application Server ND 7.0 Royal Cyber Inc. JVM related problems Application server stops responding Server crash Hung process Out of memory condition Performance degradation Check if the
Holistic Performance Analysis of J2EE Applications By Madhu Tanikella In order to identify and resolve performance problems of enterprise Java Applications and reduce the time-to-market, performance analysis
DCML Framework Use Cases Introduction Use Case 1: Monitoring Newly Provisioned Servers Use Case 2: Ensuring Accurate Asset Inventory Across Multiple Management Systems Use Case 3: Providing Standard Application
Oracle WebLogic Server 11g Administration This course is designed to provide instruction and hands-on practice in installing and configuring Oracle WebLogic Server 11g. These tasks include starting and
Chapter 2 TOPOLOGY SELECTION SYS-ED/ Computer Education Techniques, Inc. Objectives You will learn: Topology selection criteria. Perform a comparison of topology selection criteria. WebSphere component
zen Platform technical white paper The zen Platform as Strategic Business Platform The increasing use of application servers as standard paradigm for the development of business critical applications meant
CHAPTER 1 - JAVA EE OVERVIEW FOR ADMINISTRATORS Java EE Components Java EE Vendor Specifications Containers Java EE Blueprint Services JDBC Data Sources Java Naming and Directory Interface Java Message
MEASURING WORKLOAD PERFORMANCE IS THE INFRASTRUCTURE A PROBLEM? Ashutosh Shinde Performance Architect email@example.com Validating if the workload generated by the load generating tools is applied
Architectural Overview CatDV Pro Workgroup Server Square Box Systems Ltd May 2003 The CatDV Pro client application is a standalone desktop application, providing video logging and media cataloging capability
Objectives At the end of this chapter, participants will be able to understand: Web server management options provided by Network Deployment Clustered Application Servers Cluster creation and management
How to Build an E-Commerce Application using J2EE Carol McDonald Code Camp Engineer Code Camp Agenda J2EE & Blueprints Application Architecture and J2EE Blueprints E-Commerce Application Design Enterprise
A New View on Application Management www.quest.com/newview Monitoring applications in multitier environment Uroš Majcen firstname.lastname@example.org 2008 Quest Software, Inc. ALL RIGHTS RESERVED. Management Challenges
ORACLE: Oracle Press Oracle WebLogic Server 11g Administration Handbook Sam R. Alapati Mc Graw Hill New York Chicago San Francisco Lisbon London Madrid Mexico City Milan New Delhi San Juan Seoul Singapore
Continuous Availability Suite: Neverfail s Continuous Availability Suite is at the core of every Neverfail solution. It provides a comprehensive software solution for High Availability (HA) and Disaster
Stock Trader System Architecture Description Michael Stevens email@example.com http://www.mestevens.com Table of Contents 1. Purpose of Document 2 2. System Synopsis 2 3. Current Situation and Environment
S y s t e m A r c h i t e c t u r e V e r s i o n 5. 0 Page 1 Enterprise etime automates and streamlines the management, collection, and distribution of employee hours, and eliminates the use of manual
WHITEPAPER PIVOTAL CRM ARCHITECTURE Built for Enterprise Performance and Scalability WHITEPAPER PIVOTAL CRM ARCHITECTURE 2 ABOUT Performance and scalability are important considerations in any CRM selection
Load Balancing using Pramati Web Load Balancer Satyajit Chetri, Product Engineering Pramati Web Load Balancer is a software based web traffic management interceptor. Pramati Web Load Balancer offers much
White Paper: 1) Architecture Objectives: The primary objective of this architecture is to meet the following requirements (SLAs). Scalability and High Availability Modularity and Maintainability Extensibility
Web Application Hosting Cloud Architecture Executive Overview This paper describes vendor neutral best practices for hosting web applications using cloud computing. The architectural elements described
Test Run Analysis Interpretation (AI) Made Easy with OpenLoad OpenDemand Systems, Inc. Abstract / Executive Summary As Web applications and services become more complex, it becomes increasingly difficult
Oracle University Contact Us: 1.800.529.0165 Oracle WebLogic Server 11g: Administration Essentials Duration: 5 Days What you will learn This Oracle WebLogic Server 11g: Administration Essentials training
Page 1 of 9 Published on The O'Reilly Network (http://www.oreillynet.com/) http://www.oreillynet.com/pub/a/onjava/2002/07/17/tomcluster.html See this if you're having trouble printing code examples Clustering
This course teaches system/application administrators to setup, configure and manage an Oracle WebLogic Application Server, its resources and environment and the Java EE Applications running on it. This
Code:1Z0-599 Titre: Oracle WebLogic Server 12c Essentials Version: Demo http://www.it-exams.fr/ QUESTION NO: 1 You deploy more than one application to the same WebLogic container. The security is set on
Resource Utilization of Middleware Components in Embedded Systems 3 Introduction System memory, CPU, and network resources are critical to the operation and performance of any software system. These system
Performance Prediction, Sizing and Capacity Planning for Distributed E-Commerce Applications by Samuel D. Kounev (firstname.lastname@example.org) Information Technology Transfer Office Abstract Modern e-commerce
September 2015 Brocade Virtual Traffic Manager and Oracle EBS 12.1 Deployment Guide 2015 Brocade Communications Systems, Inc. All Rights Reserved. ADX, Brocade, Brocade Assurance, the B-wing symbol, DCX,
2 Apache Tomcat Load-balancing and Clustering Mark Thomas, 20 November 2014 Introduction Apache Tomcat committer since December 2003 email@example.com Tomcat 8 release manager Member of the Servlet, WebSocket
Planning the Migration of Enterprise Applications to the Cloud A Guide to Your Migration Options: Private and Public Clouds, Application Evaluation Criteria, and Application Migration Best Practices Introduction
Site24x7: Powerful, Agile, Cost-Effective IT Management from the Cloud Ensuring Optimal Performance and Quality Web Experiences Must-know facts about Site24x7: We bring expertise gained from ManageEngine
PH073-Williamson.book Page 1 Thursday, June 17, 2004 3:53 PM C H A P T E R 1 Introduction to WebSphere Administration T his book continues the series on WebSphere Application Server Version 5 by focusing
Exam : Oracle 1Z0-108 Title : Oracle WebLogic Server 10gSystem Administration Version : DEMO 1. Scenario : A single tier WebLogic cluster is configured with six Managed Servers. An Enterprise application
Learning GlassFish for Tomcat Users White Paper February 2009 Abstract There is a direct connection between the Web container technology used by developers and the performance and agility of applications.
Revised: September 2, 2010 Created: March 31, 2010 Author: Jérôme Horber CONTENTS Summary This document describes the system requirements and possible deployment architectures for MEGA Web Application.
This document describes how the Meraki Cloud Controller system enables the construction of large-scale, cost-effective wireless networks. Copyright 2009 Meraki, Inc. All rights reserved. Trademarks Meraki
Converting Java EE Applications into OSGi Applications Author: Nichole Stewart Date: Jan 27, 2011 2010 IBM Corporation THE INFORMATION CONTAINED IN THIS REPORT IS PROVIDED FOR INFORMATIONAL PURPOSES ONLY.
Base One's Rich Client Architecture Base One provides a unique approach for developing Internet-enabled applications, combining both efficiency and ease of programming through its "Rich Client" architecture.
Service Virtualization: Managing Change in a Service-Oriented Architecture Abstract Load balancers, name servers (for example, Domain Name System [DNS]), and stock brokerage services are examples of virtual
XpoLog Center Suite Data Sheet General XpoLog is a data analysis and management platform for Applications IT data. Business applications rely on a dynamic heterogeneous applications infrastructure, such
HOW TO CONFIGURE PASS-THRU PROXY FOR ORACLE APPLICATIONS Overview of Oracle JInitiator Oracle JInitiator enables users to run Oracle Forms applications using Netscape Navigator or Internet Explorer. It
DB2 Connect for NT and the Microsoft Windows NT Load Balancing Service Achieving Scalability and High Availability Abstract DB2 Connect Enterprise Edition for Windows NT provides fast and robust connectivity
Kony Mobile Application Management (MAM) Kony s Secure Mobile Application Management Feature Brief Contents What is Mobile Application Management? 3 Kony Mobile Application Management Solution Overview
Amazon Web Services Primer William Strickland COP 6938 Fall 2012 University of Central Florida AWS Overview Amazon Web Services (AWS) is a collection of varying remote computing provided by Amazon.com.
I. Basics 1. What is Application Server 2. The need for an Application Server 3. Java Application Solution Architecture 4. 3-tier architecture 5. Various commercial products in 3-tiers 6. The logic behind
VMware vcloud Automation Center 6.1 Reference Architecture T E C H N I C A L W H I T E P A P E R Table of Contents Overview... 4 What s New... 4 Initial Deployment Recommendations... 4 General Recommendations...
Veeam ONE What s New in v9? Veeam ONE is a powerful monitoring, reporting and capacity planning tool for the Veeam backup infrastructure, VMware vsphere and Microsoft Hyper-V. It helps enable Availability
SERENA SOFTWARE Scaling for the Enterprise Authors: Bill Weingarz, Pete Dohner, Kartik Raghavan, Amitav Chakravartty 2012-06-11 Table of Contents Who Should Read This Paper?... 3 Introduction... 3 Architecture...
Glassfish, JAVA EE, Servlets, JSP, EJB Java platform A Java platform comprises the JVM together with supporting class libraries. Java 2 Standard Edition (J2SE) (1999) provides core libraries for data structures,
Meet the challenges of managing composite applications IBM Tivoli Composite Application Manager for WebSphere Highlights Simplify management throughout the life cycle of complex IBM WebSphere-based J2EE
Managing SOA Course materials may not be reproduced in whole or in part without the prior written permission of IBM. 4.0.3 4.0.3 Unit objectives After completing this unit, you should be able to: Explain
Oracle BI Publisher Enterprise Cluster Deployment An Oracle White Paper August 2007 Oracle BI Publisher Enterprise INTRODUCTION This paper covers Oracle BI Publisher cluster and high availability deployment.
Mon Jan 26 2004 18:14:15 America/New_York Published on The O'Reilly Network (http://www.oreillynet.com/) http://www.oreillynet.com/pub/a/onjava/2001/09/26/load.html See this if you're having trouble printing
Using Tomcat with CA Clarity PPM April 2014 Page 2 - Revision 1.0 TOMCAT Apache Tomcat is the black-box solution that comes bundled with CA Clarity PPM. The following topics will outline the benefits our
..................................... PEPPERDATA IN MULTI-TENANT ENVIRONMENTS technical whitepaper June 2015 SUMMARY OF WHAT S WRITTEN IN THIS DOCUMENT If you are short on time and don t want to read the
19.10.11 Amazon Elastic Beanstalk A Short History of AWS Amazon started as an ECommerce startup Original architecture was restructured to be more scalable and easier to maintain Competitive pressure for
Technical White Paper: WEb Load Testing To perform as intended, today s mission-critical applications rely on highly available, stable and trusted software services. Load testing ensures that those criteria
Dependency Free Distributed Database Caching for Web Applications and Web Services Hemant Kumar Mehta School of Computer Science and IT, Devi Ahilya University Indore, India Priyesh Kanungo Patel College
Security Guide BlackBerry Enterprise Service 12 for ios, Android, and Windows Phone Version 12.0 Published: 2015-02-06 SWD-20150206130210406 Contents About this guide... 6 What is BES12?... 7 Key features
StreamServe Persuasion SP5 StreamStudio Administrator s Guide Rev B StreamServe Persuasion SP5 StreamStudio Administrator s Guide Rev B OPEN TEXT CORPORATION ALL RIGHTS RESERVED United States and other
HP Insight Management Agents architecture for Windows servers Technology brief, 2 nd edition Introduction... 3 A first look at the Insight Management Agents architecture... 3 HP Insight Management agents...
Azure Scalability Prescriptive Architecture using the Enzo Multitenant Framework Many corporations and Independent Software Vendors considering cloud computing adoption face a similar challenge: how should
Simplified Management With Hitachi Command Suite By Hitachi Data Systems April 2015 Contents Executive Summary... 2 Introduction... 3 Hitachi Command Suite v8: Key Highlights... 4 Global Storage Virtualization
Scaling Web Applications in a Cloud Environment using Resin 4.0 Abstract Resin 4.0 offers unprecedented support for deploying and scaling Java and PHP web applications in a cloud environment. This paper
Designing a Cloud Storage System End to End Cloud Storage When designing a cloud storage system, there is value in decoupling the system s archival capacity (its ability to persistently store large volumes
1. Comments on reviews a. Need to avoid just summarizing web page asks you for: i. A one or two sentence summary of the paper ii. A description of the problem they were trying to solve iii. A summary of
Enterprise Private Cloud Storage The term cloud storage seems to have acquired many definitions. At Cloud Leverage, we define cloud storage as an enterprise-class file server located in multiple geographically
IBM WebSphere Server Administration This course teaches the administration and deployment of web applications in the IBM WebSphere Application Server. Duration 24 hours Course Objectives Upon completion
Siemens PLM Connection High Availability of Teamcenter Enterprise Mark Ludwig Copyright Siemens Copyright PLM Software Siemens Inc. AG 2008. All rights reserved. Teamcenter Digital Lifecycle Management
Exam : IBM 000-217 Title : Test 217, WebSphere Commerce V6.0. Application Development Version : R6.1 Prepking - King of Computer Certification Important Information, Please Read Carefully Other Prepking
1 Portal In Anger Ray Ploski Team Lead - Solutions Architecture, JBoss by Red Hat June 22, 2010 Agenda: Location of Atlantis Video of Big Foot Guided tour of Area 52 Meaning of Life Slight Change of Plans
LinuxWorld Conference & Expo Server Farms and XML Web Services Jorgen Thelin, CapeConnect Chief Architect PJ Murray, Product Manager Cape Clear Software Objectives What aspects must a developer be aware