Grid computing and collaboration technology in support of fusion

Size: px
Start display at page:

Download "Grid computing and collaboration technology in support of fusion"

Transcription

1 Grid computing and collaboration technology in support of fusion energy sciences D.P. Schissel General Atomics, P.O. Box 85608, San Diego, California Contact Author: D.P. Schissel, General Atomics, P.O. Box 85608, San Diego, California , Phone: (858) , Fax: (858) Received ( Abstract. Science research in general and magnetic fusion research in particular continue to grow in size and complexity resulting in a concurrent growth in collaborations between experimental sites and laboratories worldwide. The simultaneous increase in wide area network speeds has made it practical to envision distributed working environments that are as productive as traditionally collocated work. In computing power, it has become reasonable to decouple production and consumption resulting in the ability to construct computing grids in a similar manner as the electrical power grid. Grid computing, the secure integration of computer systems over high speed networks to provide on-demand access to data analysis capabilities and related functions, is being deployed as an alternative to traditional resource sharing among institutions. For human interaction, advanced collaborative environments are being researched and deployed to have distributed group work that is as productive as traditional meetings. The DOE SciDAC initiative has sponsored several Collaboratory Projects, including the National Fusion Collaboratory Project, to utilize recent advances in grid computing and advanced collaborative environments to further research in several specific scientific domains. For 1

2 fusion, the collaborative technology being deployed is being used in present day research and is also scalable to future research, in particular to the ITER experiment that will require extensive collaboration capability worldwide. This paper briefly reviews the concepts of grid computing and advanced collaborative environments and gives specific examples of how these technologies are being used in fusion research today. 2

3 I. INTRODUCTION Science today is as much based on large teams of scientists as on the efforts of individual experimentalists and theorists. This shift over time to science as a large team enterprise is the result of both increasingly complex problems and the availability of increasingly powerful technology. Whether one considers the Compact Muon Solenoid experiment at the Large Hadron Collider in CERN, the design of ITER, or the Human Genome Project, large teams of geographically distributed scientists are working jointly on experimental and theoretical problems to advance their science. Certainly, the increase in computing power over time (doubling every 18 months on average) has helped to propel science forward. But during this same period, computer networks have increased in speed approximately twice as fast increasing two orders of magnitude in five years. If this trend continues, computer connectivity over wide area networks will be essentially unlimited and the need to solve computational problems locally disappears. Although not new [1,2], the concept of sharing distributed computing power has been the focus of intense computer science research over the past decade. Termed the Grid, it offers the potential for providing secure access to remote services, to allow scientific collaborators to share resources on an unprecedented scale, and for geographically distributed groups to work together in ways not previously possible [3]. This paper reviews the general concepts of Grid computing and Advanced Collaborative Environments (ACE) as they apply to scientific research. With those concepts in hand, the paper gives an overview of the National Fusion Collaboratory Project that is applying, both separately and together, Grid and ACE technologies to advance magnetic fusion science. Combined together, this new infrastructure is being molded to create the collaborative control room for tokamak operations. Although being used in today s research, these new capabilities need to be significantly enhanced before 3

4 their usage is routine. Finally, the paper concludes with a discussion of how Grids and ACE can be applied to future fusion research including ITER. 4

5 II. GRID COMPUTING The decoupling of production and consumption of food, water, and power has played a major role in the modernization of society over the past several centuries. During the time of the founding fathers each home typically had its own wood stove for heat; production and consumption were directly linked. A little over one hundred years ago, Thomas Edison established a central electrical generation station at Pearl Street in lower Manhattan. Although confined to a radius of approximately a half-mile from the station, homes and streets in that area all had the ability to use this centrally generated electricity. The consumption and production of electricity had started to be decoupled. Today, electrical consumption is completely decoupled from production and it has enabled ondemand access, the economics of scale, consumer flexibility, and new devices. It is this analogy of the electrical power grid that has given rise to the term grid for computing [3]; the large-scale integration of computer systems via high-speed networks to provide on-demand access to computational resources (data, codes, visualization) that are not available to an individual at one location. To users, the highly integrated networks that embody grid systems are transparent so that services furnished from afar appear to be provided by local computers. Returning to the analogy of the electrical power grid, when a user plugs a device into an electrical wall socket, the how and where of electrical generation are immaterial. Useful functions (electricity) are hidden by an interface (plug) that conceals the details (power plant) of how they are implemented allowing the individual to concentrate on using the function. In a traditional computing environment, software users typically install and run programs on a local machine (CPU cycle production and consumption are directly linked). This requires that developers create and maintain versions of their software for the different platforms used by their users. This also requires users to update their local installations as the software is updated. In a grid computing environment applications, 5

6 systems, and other computing resources are abstracted into services [4]. This abstraction allows users to invoke services on local or remote hosts without concerning themselves with the details of how such services are implemented. Utilizing remote hosts implies the user has transitioned out of their own administrative domain into that of another organization. To be effective, this type of sharing needs to be governed by a set of rules including what is shared, who is allowed to share, and under what conditions sharing can occur. When physically different administrative domains work together in such a coordinated fashion with a set of clearly defined sharing rules they form what is called a Virtual Organization (VO) [3]. It is the VO that allows geographically separated groups to share computer resources in a controlled fashion to work toward a common goal (Fig. 1). A. Security The Internet is an open system, where the identity of the communicating partners is not easy to assure. Furthermore, the communication path traverses an indeterminate set of routing hosts and may include any number of eavesdropping and active interference possibilities. Thus, Internet communication is much like anonymous postcards, which are answered by anonymous recipients. The ability to remove local administrative boundaries to form VOs implies the ability to have a sufficient security infrastructure to ensure that the sharing rules that define the VO can be enforced. Thus a major component for the successful implementation of computer grids is security [5] including authentication, authorization, data encryption and so on. For this paper and understanding of basic grid security, we will cover only the first two topics. Authentication is the process in a computerized transaction that gives assurance that a person or computer acting on a person s behalf is not an impostor. Authorization is the process of determining, by evaluating applicable access control information, whether a 6

7 subject is allowed to have the specified types of access to particular resource. Once a subject is authenticated, it may be authorized to perform different types of access. To use an analogy from everyday life, when boarding a commercial aircraft flight, a drivers license provides authentication that you are who you say you are while the airplane boarding pass determines that you have the right to use the requested resource (airplane). Unlike the physical interaction that occurs in the above analogy with the driver s license, there is no physical interaction in a computer transaction and therefore identity verification is harder. Two parties can communicate securely by using symmetric cryptography [6]. In this system, two parties (Bob and Patrick) agree on a cryptosystem and also agree on a key. For Bob to send a message to Patrick, he first encrypts the message using the key, sends this ciphertext message to Patrick, who then uses the same key to decrypt the message so it can be read. The security for this system rests in the key, divulging the key means that anyone can encrypt or decrypt messages. The analogy is placing a message in a safe where the key is the combination and anyone with the combination can open the safe. Assuming a separate key is used for each pair of users in a network, the total number of keys increases rapidly as the number of users increases; for n users there are nn ( 1) 2 keys. For a worldwide encryption system, securing this number of keys can be a daunting task. Public-key cryptograph [7] presents an easier to manage solution for key distribution that scales to large groups. This system uses two different keys, one public and one private, where it is computationally hard to deduce the private key from the public key. Anyone with the public key can encrypt a message but not decrypt it and only the person with the private key can decrypt the message. Mathematically the process is based on the trap-door one-way function [8], which are relatively easy to compute but significantly harder to reverse unless the secret is known. That is, given x it is easy to compute f( x), but given f( x) it is hard to compute x. However, there is some secret information y, such that given f( x) and y it is easy to compute x. Using this system, Bob sends a 7

8 message to Patrick encrypting it using Patrick s public-key. Patrick then uses his privatekey to decrypt the message so it can be read. If Patrick places his public-key in a publicly available database, Bob can send a secure message with no prior communication between the two (Fig. 2). Therefore, public-key cryptography solves the key-management problem of symmetric cryptosystems. Public Key Infrastructure (PKI) is a technology to distribute and use asymmetrical keys. PKI gives trust that the public key being used truly belongs to the person or machine with whom/which they wish to communicate [9]. In the previous example, trust needs to be established for Bob to believe that the public-key he uses really does belong to Patrick instead of an impostor. This trust is established through the usage of certificate authorities (CAs) who issue X.509 certificates where a unique identity name and the public key of an entity are bound together through the digital signature of that CA (certificate = trust + public-key). Typically, a registration authority (RA) is responsible for the identification and authentication of certificate subscribers before the CA issues certificates. Once a user s identity has been validated they are still not given open access to all resources (codes, computers, visualization tools or data). These are made available only to those users who have the proper authorization. Presently there exists no standard for authorization on grids. However it is often broken down into a policy decision point (PDP) and a policy enforcement point (PEP). The PEP is a software component, wrapped around a resource, that either allows or denies access to a resource. The PDP, usually one central authority accessed by all PEP requests, decides what resource access is allowed (Fig. 3). The implementation for establishing the identity of a consumer of a resource (authentication) and for determining whether an operation is consistent with agreed upon sharing rules (authorization) frames and ultimately defines the virtual organization. We 8

9 will now turn to how grids can be used by VOs by examining data management, computing, and visualization resources. B. Data management Although data has been stored in different methodologies throughout human history [10], it is the relational model for computer data storage where both entities and relationships are represented in a uniform way [11] that is most commonly used today. The relational database with its Structured Query Language (SQL) is well suited to the client-server model with a graphical user interface. In its simplest form, data management consists of one administrative domain and one physical data location. As data collections have grown, data under one administrative domain has become distributed over numerous physical locations. Today, data management on grids involves data spread over many administrative domains and many physical locations that taken together comprise the virtual organization discussed previously. Data management on grids has many challenges including diverse usage scenarios, heterogeneity at all system levels, and performance demands associated with access, manipulation, and analysis of large quantities of data. To be useful to the scientific community, users must be able to discover desired data based on metadata attributes. Metadata is the information about data describing, for example, the content, quality, condition, and other characteristics of the data. Once found, data needs to be efficiently moved between storage locations or between programs and storage. To be efficient, data movement includes replication, caching, and bulk data access. Finally, the coupling of computations with operations on data resources introduces new optimization problems but this functionality becomes critical as data repositories grow in size. An example of efficient data access in a grid environment is the file oriented access supported by GridFTP [12] that provides a uniform interface to various storage systems. GridFTP supports parallel data transfers as well as mechanisms for reliable and 9

10 restartable data transfer that is critical for very large data collections. An example of a large-scale science project utilizing GridFTP along with a number of other technologies is the Earth Systems Grid project [13] that is serving multi-terabits of data to climate researchers worldwide. C. Computing For computing to occur in a grid environment, mechanisms must be provided for starting programs and monitoring and controlling the execution of the resulting processes. Useful also are management mechanisms that allow for control over allocated resources as well advanced reservation capabilities. Informational discovery is required to obtain information about the structure and state of a computational resource (e.g. load, memory). The Globus toolkit [14] is an example of a software package that is open source and contains a set of services and software libraries that support grids and grid applications. It has become the foundation for many grid projects worldwide in both academia and industry. An example of such a project was EUROGRID [15] that connected major academic and commercial centers in Europe with an emphasis on high-performance applications using specialized architectures. This project demonstrated distributed simulation codes from different application areas including biomolecular, weather prediction, structural analysis, and real time data processing. D. Visualization There are a number of approaches that can be undertaken to support visualization of very large data sets in a grid environment. At one end, parallel rendering [16] can be performed on nodes spread throughout the virtual organization to support graphical discovery at one location. The other end of the spectrum would be to move the data to the scientist and do everything locally. Where in the spectrum of solutions the answer lies is 10

11 typically very dependent on the problem being solved and the computational constraints imposed by the existing infrastructure (data set size, interactivity rates, local visualization capability, etc.). Further, a grid based visualization system will most likely need to be dynamic as a user might need to interact with the system, for example to trade visualization accuracy for frame rate. One example of such a visualization system is ParaView [17], being developed as an open source initiative to develop a multi-platform visualization application to support distributed computational models to process large datasets. The TeraGrid Project [18], is deploying a distributed infrastructure for open scientific research including 20 teraflops of computing power distributed at nine sites, 1 petabyte of data storage, grid computing toolkits, and high-resolution visualization environments. ParaView is being deployed on TeraGrid in an attempt to solve the problem of grid based visualization of very large datasets. 11

12 III. ADVANCED COLLABORATIVE ENVIRONMENTS Researchers often want to aggregate not only data and computing power, but also human expertise. Collaborative problem formulation, data analysis, and the like are important grid applications. The goals of the advanced collaborative environment (ACE) is to use computer mediated communications techniques to enhance work environments, to enable increased productivity for collaborative work, and to exploit the use of highperformance computing technologies to improve the effectiveness of large-scale collaborative work environments. To be effective, collaboration environments should provide lightweight and ubiquitous components that support a wide variety of interaction modes. Such remote work can range from the highly structured (e.g. formal presentations) to more informal, spontaneous collaborations (e.g. cooperative software development). Many of today s scientific collaboratory tools such as videoconferencing tools are highly interactive and only support formal meetings well. Although videoconferencing is an important part of a collaborative environment, much of the work of scientific collaboration requires more informal and asynchronous mechanisms (important for world-wide collaborations). Engaging in informal interactions and sharing documents and data have been shown to be an important part of an effective collaboration [19]. The most fundamental characteristic of a collaborative environment is ubiquity. Collaborators should be able to enter and work within the environment from their desktop machine, their laptop, another user s computer, or any other digital device, or from any location. The collaboration tools need to provide a real benefit to all the users instead of being only one-sided, which is counterproductive [20]. Many tools have been developed to facilitate remote collaboration and a complete review is beyond the scope of this paper. Text-based messaging systems such as America On-line Instant Messaging (AIM), ICQ, Yahoo Messenger, and Jabber are primarily 12

13 intended for one-on-one conversation but they have been recently extended to include group-chat as well as file transfer. Several systems such as the Virtual Room Videoconferencing System (VRVS) [21] and WebEx offer web-oriented, low-cost, bandwidth efficient, extensible means of videoconferencing and remote collaboration over IP networks. VRVS is used extensively in the High Energy Physics field and creates a virtual meeting room where participants gather as if they were together in the same physical room. VRVS transmits all active video and audio channels to all participants via a network of reflectors. Integrated into VRVS is a chat capability that allows for back-channel communication during the virtual meeting. The Mbone videoconferencing tools (vic, vat, rat, wb, and sdr) provide multicast, multi-way videoconferencing over IP networks that allows all users to be seen and heard as equal participants. The Access Grid software [22] extends the Mbone work to include the ability to utilize for scientific research a complex multi site visual and collaborative experience integrated with high end visualization environments (Fig. 4). AG nodes range in size from the individual desktop similar to VRVS to a very large meeting room. Integrated into the AG system is a modified VNC (Virtual Network Computing) that allows for more efficient interactive sharing of complex visualizations. VNC [23] allows Internet sharing of a computer desktop with one or more remote clients. Within the AG environment, VNC is being used to interactively share applications during working meeting and to broadcast electronic slide presentation. Since its inception, VRVS has been extended to also interact with Mbone and Access Grid systems as well as the integration of VNC to create a more interactive virtual meeting. Tiled display walls [24] are being investigated to enhance the collaborative work environment of large groups of collocated individuals. A tiled display wall utilizes multiple projectors tiled together to build a bright, high-resolution, seamless display with 16 ft x 8 ft, 20 million pixel displays not uncommon (Fig. 5). With the increased speeds of computers, networks, and graphics cards, the ability to deploy tiled displays at low 13

14 cost with commodity components is being realized. Such a display offers a large-format environment for presenting high-resolution visualizations or multi-source smaller visualizations to a collaborative group than would be possible on standard displays. For collocated individuals, this interactive shared visualization takes the place of passing around a graphical printout or calling over scientists to collaboratively view a normal desktop display. Given the ability to share visualizations from remote locations to tiled displays, off-site scientists can interactively share visualizations and participate in largegroup discussions, something previously not possible. The computer science research being conducted within the area of tiled displays includes parallel rendering, user interfaces, image blending, computational alignment, and color balancing [25]. 14

15 IV. THE NATIONAL FUSION COLLABORATORY PROJECT Historically, efforts to improve collaboration within the U.S. fusion community have included sharing of resources and co-development of tools mostly carried out on an ad hoc basis. The community has considerable experience in placing remote collaboration tools into the hands of real users [26]. The ability to remotely view operations and to control selected instrumentation and analysis tasks was demonstrated as early as 1992 [27]. Full remote operation of an entire tokamak experiment was tested in 1996 [28,29]. The National Fusion Collaboratory Project [30] is funded by the United States Department of Energy (DOE) under the Scientific Discovery through Advanced Computing Program (SciDAC) to develop a persistent infrastructure to enable scientific collaboration for all aspects of magnetic fusion research. Initiated in late 2001, this project builds on the past collaborative work performed within the U.S. fusion community and adds the component of computer science research done within the USDOE Office of Science, Office of Advanced Scientific Computer Research. The project is a collaboration itself uniting fusion scientists and computer scientists from seven institutions to form a coordinated team. This group is leveraging existing computer science technology where possible and extending and/or creating new capabilities where required. The vision for FusionGrid is that experimental and simulation data, computer codes, analysis routines, visualization tools, and remote collaboration tools are to be thought of as network services which represents a fundamental paradigm shift for the fusion community. In this model, an application service provider (ASP) provides and maintains software resources as well as the necessary hardware resources. The project is creating a robust, user friendly collaborative software environment and making it available to the more than one thousand fusion scientists in forty institutions who perform magnetic fusion research in the United States. In particular, the project is developing and deploying 15

16 a national Fusion Energy Sciences Grid (FusionGrid) that is a system for secure sharing of computation, visualization, and data resources over the Internet. The FusionGrid goal is to allow scientists at remote sites to fully participate in experimental and computational activities as if they were working at a common site thereby creating a virtual organization of the U.S. Fusion community. This Grid s resources are protected by a shared security infrastructure including strong authentication to identify users and fine-grain authorization to allow stakeholders to control their own resources. FusionGrid will shield the users from software implementation details and allow a sharper focus on the physics with transparency and ease-of-use being the crucial elements. In this environment, access to services is stressed rather than data or software portability. FusionGrid is not focused on computer cycle scavenging (e.g. SETI@home) or distributed supercomputing that are typical justifications for Grid computing, but simply on making the ASP paradigm effective for Grids. Accomplishing the Project s goals will advance scientific understanding and innovation in magnetic fusion research by enabling more efficient use of existing experimental facilities and more effective integration of experiment, theory, and modeling. Physics productivity will be increased by (1) allowing more transparent and uniform access to analysis and simulation codes, to data, and to visualization tools resulting in more researchers having access to more resources, (2) creating a standard tool set for remote data access, security, and visualization allowing more researchers to build these into their own services, (3) enabling more efficient utilization of experimental time through more powerful between pulse data analysis and through enhanced human participation (both remote and collocated) resulting in faster experimental progress at less cost, (4) facilitating the comparison of theory and experiment through transparent remote data access with appropriate security, and (5) facilitating multi-institution collaborations through the creation of the standard toolset. The Project will also increase the productivity of code and tool developers by (1) supporting more users with fewer 16

17 installations at reduced cost (e.g. see Section IV.B TRANSP), (2) facilitating shared code development projects resulting in more rapid code creation through enhanced interaction with remote staff, code sources, data and visualizations (e.g. shared code debugging), and (3) creating a standard tool set for remote data access, security, and visualization allowing these services to be easily built into new code systems. A. FusionGrid security FusionGrid security employs Public Key Infrastructure (PKI) to secure communication on the Internet through the use of a public and private cryptographic key pair that is obtained and shared through a trusted authority as was discussed previously. FusionGrid uses the X.509 certificate standard and the FusionGrid CA to implement PKI for secure communication. A scientist who desires to join FusionGrid will generate a public/private key pair and apply to the FusionGrid CA for an X.509 certificate as discussed in Section II.A. That request goes to the RA who will verify their identity and the validity of their request (i.e. determine they are known members of the community and they have a reason to join FusionGrid). FusionGrid certificates are managed on behalf of the user by a myproxy online certificate repository [31] securely installed at LBNL. In this system, the user s long-term certificate (private key+trust) is securely stored on the myproxy server and access to FusionGrid is accomplished by having the user submit their username and password to the myproxy server. The submittal process is typical done behind the scene on behalf of the user; they only have to type their username and password. With this FusionGrid login, the myproxy server issues a short-term certificate that is then used for authentication. By storing the users long-term certificates on the myproxy server, users no longer have to manage their certificate, but instead delegate that task to a FusionGrid administrator. 17

18 The secure authenticated connections are accomplished using the Globus Toolkit [14]. All a user needs to do is submit their username and password to the myproxy server once per day. This single sign-on is accomplished behind the scenes by the use of a short-lived proxy certificate that is derived from the user s long-term X.509 certificate. The proxy certificate uses its own unencrypted private key, so that it can make frequent authenticated connections on behalf of the user to multiple services without requiring additional password interactions with the user. The benefit to the user is that they need only log-on once no matter how many different services they desire to use. Centralized authorization of FusionGrid resources is accomplished through the Resource Oriented Authorization Management System (ROAM). This system allows a resource provider to implement either a simple or complex authorization policy using a web browser interface. System flexibility is maintained since the resource provider is allowed to either use existing permission levels or define their own as required. The system is implemented using an Apache-based web server and a single PHP (a recursive acronym for PHP: Hypertext Preprocessor) script. PHP is an open-source, server-side HTML embedded scripting language used to create dynamic Web pages (e.g. search results from a database). A dynamic Web page is a page that interacts with the user, so that each user visiting the page sees customized information. A PostgreSQL database is used to manage all the authorization information. Access to this information is done via secure HyperText Transport Protocol (HTTPS) using either the user s certificate, if present, or by a myproxy login. Resources also check for authorization using HTTPS communication. B. FusionGrid data and computing Data access on FusionGrid has been made available using the MDSPlus data acquisition and data management system [32] combined with the relational database 18

19 Microsoft SQL server. MDSplus, developed jointly by MIT, LANL, and the IGI in Padua, Italy, is by far the most widely used data system in the international fusion program. Based on a client/server model, MDSplus provides a hierarchical, selfdescriptive structure for simple and complex data types [33,34] and is currently installed and used in a variety of ways by about 30 experiments, spread over 4 continents. It is deployed as a complete data acquisition and analysis systems for C-Mod (MIT); RFX (IGI, Padua); TCV (EPFL, Switzerland); NSTX (PPPL); Heliac (ANU, Australia); MST (U. Wisconsin); HIT (U. Washington); CHS (NIFS, Japan); and LDX (MIT). It is used to store processed data for DIII-D, for the collaborative data archives assembled by the ITPA, and for the inputs and outputs of several widely used codes including EFIT, TRANSP, NIMROD and GS2. JET and ASDEX-Upgrade are using MDSplus as a remote interface to existing data stores and KSTAR has adopted it as a data acquisition engine for data stored in other formats. The result is a de facto standard that greatly facilitates data sharing and collaborations across institutions. MDSplus and the Globus Toolkit have been combined to create secure X.509 certificate based client/server data access on FusionGrid using the standard MDSplus interface without any loss in speed or functionality. SQL Server is securely accessible via MDSplus since a production of release of Globus for Windows is not available. Presently, the three main MDSplus experimental data repositories at Alcator C-Mod, DIII-D, and NSTX are securely available on FusionGrid. Data management by MDSplus of large datasets generated by simulation codes is presently being tested using results from NIMROD simulations. NIMROD is a 3D MHD simulation code that runs on very large parallel computers. Using the MDSplus server at DIII-D, output from NIMROD runs up to 100 GB have been stored and served to users for further data analysis and visualization. Although successful, this storage methodology proved to be inefficient. The installation of an MDSplus server on the NERSC LAN along with the highperformance computational servers has been undertaken to investigate increased 19

20 throughput capability. Parallel network data transport are also being investigated in order to overcome TCP/IP flow control limits for high bandwidth, high latency connections. The code TRANSP, used for time dependent analysis and simulation of tokamak plasmas, was released as a service on FusionGrid late in 2002 [35] along with supporting infrastructure development (data storage, monitoring, user GUI) [36]. This FusionGrid service has been so successful that it has become the production system for TRANSP usage in the United States and is starting to be adopted internationally. Running on a Linux cluster at PPPL, over 4600 TRANSP runs from ten different experimental machines have been completed within the FusionGrid infrastructure (Fig. 6). European scientists use TRANSP on FusionGrid with approximately 40% of the runs performing analysis on data from European machines. This approach has drastically reduced the efforts to support and maintain the code which were previously required of the developers and by users sites. When users request TRANSP Grid services, their proxy certificate is used to verify their identity through the Globus GSI. Once authenticated, users are authorized to run TRANSP via ROAM. To use the TRANSP FusionGrid service, the inputs and outputs are stored in MDSplus trees. Sites without their own MDSplus server can also receive TRANSP output in the traditional NetCDF file format via GridFTP; a PPPL-provided script makes this GridFTP task very simple. The IDL-based PreTRANSP application is one technique to simplify TRANSP usage by assisting the user in preparing TRANSP inputs, managing code runs, and launching TRANSP. For now at least, PreTRANSP is only used for the preparation of TRANSP runs. Recently the GATO ideal MHD stability code was released as a FusionGrid computational service running on a Linux computer at General Atomics (GA). Following the same design as the TRANSP service, the time required to deploy GATO on FusionGrid was minimal. This result has given confidence that the design of FusionGrid will scale to the deployment of many services. 20

21 It is important to note that users executing a code run on either the PPPL or GA Linux systems need not deal with local computer accounts. For example, when a collaborator runs TRANSP on FusionGrid, connections made by their proxy are mapped to PPPLassigned run production accounts created specially for the TRANSP service. These run production accounts are implemented as local UNIX accounts on the PPPL cluster, and are used to ensure data privacy. Users never need to learn a new set of passwords or host names as this account mapping happens behind the scenes. This greatly simplifies the task of account administration. With multiple applications distributed throughout a Grid infrastructure, it becomes a challenge to monitor the progress and state of each application. Users of a Grid environment need to know the specific state of code runs, when their data results are available, or if the requested application is even available. To track and monitor applications on the FusionGrid, the FusionGrid Monitor (FGM) has been developed as a Java Servlet which can accept and monitor information from remote and distributed applications [36]. Currently, FGM tracks TRANSP and GATO analysis runs on the National Fusion Grid, and provides updated information on each individual run, including: current state, cpu time, wall time, comments, and access to log files that have been produced by the analysis. The Fusion Grid Monitoring system has built to provide user output through HTML, utilizing both server push and client pull capabilities. This allows multiple users to connect to FGM, view their code runs by using a web browser, and obtain updated information without excessive user input or client software. Designed in the Java language, the monitoring system is portable, and with the inclusion of the Java Expert System Shell (JESS), the system is also expandable and customizable. Online access to log files is available through FGM, utilizing anonymous FTP. FGM has been recently extended to include an Internet-accessible Java-based graphical monitoring tool, ElVis, to display results from remote simulations as they are computed. The ElVis monitoring not only shows that the remote computational service is 21

22 operating; it also allows select results to be made available in the control room or at collaborator sites even before the run is completed. B FusionGrid advanced collaborative environment The goals of FusionGrid s advanced collaborative environment service is to use computer mediated communications techniques to enhance work environments, to enable increased productivity for collaborative work, and to exploit the use of high-performance computing technologies to improve the effectiveness of large scale collaborative work environments. Examples of such collaboration include off site support of experimental operations, large group collaborations in a tokamak control room, simulation/experimental data analysis meetings, and shared code debugging. Tiled display walls are being used to enhance the collaborative work environment of the tokamak control room. Such a display offers a large-format environment for presenting high-resolution visualizations or multi-source smaller visualizations to a collaborative group than would be possible on standard displays. As a prototype FusionGrid service, tiled display walls have been tested in a variety of usage modalities ranging from two tiled walls geographically separated being tied together by software to form shared collaborative displays to a single tiled wall used for collocated group sharing and discussion. Based on the success of these tests, a 2-tile front projection system has been installed in the NSTX control room and a 3-tile rear projection system in the DIII-D control rooms (Fig. 7). Based on VNC, this service allows any researcher either in the control room or off site, with proper authentication and authorization, to share any X windows based visualization, with the entire control room. For scientists within the control room, this interactive shared visualization takes the place of passing around a graphical printout or calling over scientists to collaboratively view a normal desktop display. For scientists off site, this service gives 22

23 them the capability to interactively share visualizations and participate in experiments, something previously not possible. The software has been designed so that the remote scientist need not purchase any special hardware and they can therefore share pieces of the larger control room display wall on their single desktop display. The Access Grid is used by FusionGrid to create a service that enables group-togroup interaction and collaboration that improves the user experience significantly beyond teleconferencing. The Access Grid includes the ability to utilize for scientific research a complex multi site visual and collaborative experience integrated with highend visualization environments. Developed exclusively for a FusionGrid service, the personal interface to the Access Grid (PIG) has been developed as a low cost alternative to a full-blown conference room size Access Grid node (Fig. 8). 23

24 V. THE COLLABORATIVE CONTROL ROOM The combination of Grid computing with collaboration technologies such as the Access Grid (AG) with application sharing has the potential to dramatically improve the efficiency of experimental sciences. The combination of these technologies into a unified scientific research environment called the collaborative control room poses unique challenges but creates the possibilities of high reward in the form of increased efficiency of experiments. Magnetic fusion experiments operate in a pulsed mode. In any given day, plasma pulses are taken with approximately 10 to 20 minutes between each ~10-second pulse. For every plasma pulse, up to 10,000 separate measurements versus time are acquired at sample rates from khz to MHz, representing about a gigabyte of data. Throughout the experimental session, hardware/software plasma control adjustments are made as required by the experimental science. These adjustments are debated and discussed among the experimental team. Decisions for changes to the next pulse are informed by data analysis conducted within the roughly 20-minute between-pulse interval. Data analysis to support experimental operations includes between pulse analysis of raw acquired data as well as the merging of numerous data sources for whole-device simulation of the experimental plasma. Results of more detailed, computationally demanding predictive simulations, carried out during the planning phase prior to the experiment, are made available for comparison to the actual experimental results in real time. This mode of operation places a large premium on rapid data analysis that can be assimilated in near-real time. The experimental science can be made more efficient by pushing the boundaries in two directions. First, by running codes on geographically dispersed resources the amount and detail of both analysis and simulation results can be 24

25 increased. Second, by bringing in expertise from geographically remote teams of experts, the depth of interpretation can be increased leading to improved assimilation of those results. In order to be fully functional, the collaborative control room requires (1) secured computational services that can be scheduled as required, (2) the ability to rapidly compare experimental data with simulation results, (3) a means to easily share individual results with the group by moving application windows to a shared display, and (4) the ability for remote scientists to be fully engaged in experimental operations through shared audio, video, and applications. A. Prototype implementation A prototype implementation of the collaborative control room was developed and a simulation was demonstrated at the SC2003 meeting. The demonstration involved remote codes, resources, and scientific teams in the experiment. Offsite collaborators (SC2003 show floor, Phoenix) joined in a mockup of a DIII-D experiment located in San Diego. AG technology allowed for shared audio and video as well as shared applications. The Access Grid was used to give the remote scientist the feeling of being part of the control room at a distance by allowing the remote scientist to talk to and see the DIII-D control room, enabling the remote scientist to ask questions of the operators there as well as see what was going on in the control room as it was happening. This could never be achieved with just a telephone call. As the number of offsite collaborators grows, this can be achieved only in a limited fashion with present day videoconferencing technology. Additionally, since AG technology is open source, it can be easily expanded to add services and tools specific to a tokamak control room something that would be much more difficult, if not impossible, with commercial videoconferencing equipment. The offsite collaborators could hear DIII-D announcements from both the scientist and engineer in charge, as well as see via a Web interface the state of the pulse cycle and the 25

26 status of data acquisition and between pulse data analysis, and how much time was left before the next pulse. As the data was gathered into MDSPlus, the remote scientist was able to open standard data processing and viewing applications such as ReviewPlus and EFITViewer to start the analyzing process. Once the remote scientist identified data points of interest, they were able to move the application to a region that was shared between the Access Grid node and the control room. This area could be seen and interacted with by both parties. This kind of interaction is a significant step forward from the present situation. A modified VNC was used to handle the remote desktop sharing. Between-pulse data analysis of the plasma shape (EFIT running at PPPL) was conducted on FusionGrid through a computational reservation system that guaranteed a specific analysis to be completed within a set time window (Sec. V.B). Additionally, the TRANSP service was run at PPPL for the first time between pulses, giving the scientists data that was previously available only after the experimental day had ended. The offsite team members were able to collaborate more efficiently by being able to share their personal display with the room s shared display. This capability allowed visualizations to be efficiently compared for debate before reporting results back to the DIII-D control room. The results of this demonstration and the feedback from fusion scientists has helped sharpen the requirements for a truly collaborative control room for fusion experiments. B. Computational reservations for between-pulse analysis For the SC03 demonstration, agreement-based interactions utilizing the Globus Toolkit 3 (GT3) enabled fusion scientists to negotiate end-to-end guarantees on execution of remote codes between the experimental pulses [37]. This mode of interaction has high potential for resolving problems of provisioning in Grid computing with specifications 26

27 being drafted by the Grid Resource Allocation Agreement Protocol (GRAAP) working group of the GGF. In the demonstration, the FusionGrid TRANSP service was run on the PPPL cluster to support DIII-D operations and the results were simultaneously visualized in Phoenix and San Diego. In preparation for the demonstration, significant work was done to reduce TRANSP run production time, through both software and hardware changes, to about 6 minutes, which was found to be acceptable for an experimental run. The actual TRANSP run execution time was slightly over 3 minutes; the balance of the time was due to network data transfers. These data transfer delays will be reduced through further optimization of the software. In this first between pulse data analysis using TRANSP, only one timeslice of the experimental data was run. In principle it would be best to run a fully time-dependent TRANSP simulation. At some point in the future if the TRANSP code is parallelized, this reservation system will scale to large runs on multi-node computational clusters. C. Usage during experimental operations With the success of the prototype demonstration at the SC2003 meeting, the first usage of the collaborative control room tools during actual experimental operations was in early In this case, Dr. J.S. degrassie was the Scientific Coordinator for an experiment on the EFDA-JET fusion experiment in England from his home laboratory in San Diego (Fig. 9). Access Grid technology was used for multiple video images and a unified audio stream between San Diego and the JET control room. Web based displays gave Dr. degrassie real-time tokamak information as well as pulse cycle and general data analysis status. Although considered a success for a first time usage in actual experimental operations, the experience nevertheless indicated areas of where improvements are needed. The biggest area needing improvement was the ability to more efficiently share visualizations of data analysis. This would allow each site to look over 27

28 each others shoulder and discuss their analysis in real time. Instead, a site would verbally explain their analysis, wait for the other site to catch up by creating their own visualization, and then have the data discussion. In the time demanding environment of a tokamak control room this type of interaction is too inefficient. Subsequent to the remote collaboration with JET, similar collaborations have been undertaken with Germany (IPP) and Japan (JAERI). At times, these have been as successful as the original JET experience and at other times scientists have reverted to the telephone when the quality of the IP based audio/video service is too varied. These experiences all reinforce how low the tolerance is for error in a tokamak control room. If a remote meeting experiences audio or video difficulty, delaying or rescheduling is inconvenient but an option. In the control room, every ~20 minutes wasted means a lost opportunity for new data, and with experimental time so precious, it is unacceptable. In recognition of this situation, the NFC project has created the basic tiled wall and computing infrastructure of a control room in a separate laboratory. This setup will allow software to be further tested in a mock control room situation as well as allow for training of scientific staff before used in actual experimental operation. 28

29 VI. CONCLUSIONS AND FUTURE PROSPECTS There is a significant worldwide effort to develop and deploy Grid computing and advanced collaborative environments in support of numerous scientific disciplines including fusion energy research. For fusion energy sciences, although substantial progress has been made, more work is clearly required to reach the point where off-site participation is as rewarding as on-site participation. With the worldwide focus on ITER as the next generation machine, its success requires advanced remote collaboration capability. This capability and success needs to include more than just the experimental physics program. The final design, engineering, and construction phases will be worldwide collaborations as well and although they will not need the collaborative control room they will need the ability to richly interact with their distant colleagues. The ability to interactively share engineering drawings, conduct design reviews, and view 3-D mockups of machine components are all clearly required. Imagine an electronic tabletop display (Hitachi recently announced such a product), one can envision a drawing rolledout electronically on a table. The designer picks up her pen and begins pointing to different areas of the drawing as the ad-hoc meeting begins during ITER construction to solve an unforeseen problem. Her counterparts spread around the world have a similar table, see the drawing and her pointing, and when they look up into the accompanying display device they see images of their colleagues and hear their words. When ITER operation begins, the collaborative control room will allow scientists to share data and knowledge as readily as their engineering counterparts. Large tiled displays will present information to the assembled team in the control room. Off-site scientists will be sharing the results of their analysis to the large display as well as to individual small displays of their colleagues. ITER s integrated data acquisition and data management system not only allows for simultaneous data availability worldwide, but also the automatic starting of many data analysis tasks. Utilizing computational Grids, 29

A DISCUSSION ON DATA ACQUISITION, DATA MANAGEMENT AND REMOTE PARTICIPATION FOR ITER

A DISCUSSION ON DATA ACQUISITION, DATA MANAGEMENT AND REMOTE PARTICIPATION FOR ITER A DISCUSSION ON DATA ACQUISITION, DATA MANAGEMENT AND REMOTE PARTICIPATION FOR ITER T.W. Fredian, 1 M.J. Greenwald, 1 D.C. McCune, 2 D.P. Schissel, 3 J. Stillerman 1 1 Massachusetts Institute of Technology,

More information

DATA MANAGEMENT, CODE DEPLOYMENT, AND SCIENTIFIC VISUALLIZATION TO ENHANCE SCIENTIFIC DISCOVERY IN FUSION RESEARCH THROUGH ADVANCED COMPUTING

DATA MANAGEMENT, CODE DEPLOYMENT, AND SCIENTIFIC VISUALLIZATION TO ENHANCE SCIENTIFIC DISCOVERY IN FUSION RESEARCH THROUGH ADVANCED COMPUTING DATA MANAGEMENT, CODE DEPLOYMENT, AND SCIENTIFIC VISUALLIZATION TO ENHANCE SCIENTIFIC DISCOVERY IN FUSION RESEARCH THROUGH ADVANCED COMPUTING D.P. Schissel, 1 A. Finkelstein, 2 I.T. Foster, 3 T.W. Fredian,

More information

A GENERAL PURPOSE DATA ANALYSIS MONITORING SYSTEM WITH CASE STUDIES FROM THE NATIONAL FUSION GRID AND THE DIII D MDSPLUS BETWEEN PULSE ANALYSIS SYSTEM

A GENERAL PURPOSE DATA ANALYSIS MONITORING SYSTEM WITH CASE STUDIES FROM THE NATIONAL FUSION GRID AND THE DIII D MDSPLUS BETWEEN PULSE ANALYSIS SYSTEM A GENERAL PURPOSE DATA ANALYSIS MONITORING SYSTEM WITH CASE STUDIES FROM THE NATIONAL FUSION GRID AND THE DIII D MDSPLUS BETWEEN PULSE ANALYSIS SYSTEM S.M. Flanagan *, J.R. Burruss, C. Ludescher, a D.C.

More information

Secure Federated Light-weight Web Portals for FusionGrid

Secure Federated Light-weight Web Portals for FusionGrid Secure Federated Light-weight Web Portals for FusionGrid By: D. Aswath, M. Thompson, M. Goode, X. Lee, N. Y. Kim Presented by: Dipti Aswath GCE Workshop 2006 Second International Workshop on Grid Computing

More information

THE NATIONAL FUSION COLLABORATORY PROJECT: APPLYING GRID TECHNOLOGY FOR MAGNETIC FUSION RESEARCH

THE NATIONAL FUSION COLLABORATORY PROJECT: APPLYING GRID TECHNOLOGY FOR MAGNETIC FUSION RESEARCH THE NATIONAL FUSION COLLABORATORY PROJECT: APPLYING GRID TECHNOLOGY FOR MAGNETIC FUSION RESEARCH D.P. Schissel, 1 K. Keahey, 4 T. Araki, 4 J.R. Burruss, 1 E. Feibush, 2 S.M. Flanagan, 1 T.W. Fredian, 3

More information

Shared Display Wall Based Collaboration Environment in the Control Room of the DIII-D National Fusion Facility

Shared Display Wall Based Collaboration Environment in the Control Room of the DIII-D National Fusion Facility Shared Display Wall Based Collaboration Environment in the G. Abla a, G. Wallace b, D.P. Schissel a, S.M. Flanagan a, Q. Peng a, and J.R. Burruss a a General Atomics, P.O. Box 85608, San Diego, California

More information

A Survey Study on Monitoring Service for Grid

A Survey Study on Monitoring Service for Grid A Survey Study on Monitoring Service for Grid Erkang You erkyou@indiana.edu ABSTRACT Grid is a distributed system that integrates heterogeneous systems into a single transparent computer, aiming to provide

More information

Base One's Rich Client Architecture

Base One's Rich Client Architecture Base One's Rich Client Architecture Base One provides a unique approach for developing Internet-enabled applications, combining both efficiency and ease of programming through its "Rich Client" architecture.

More information

GA A23162 ENHANCED COMPUTATIONAL INFRASTRUCTURE FOR DATA ANALYSIS AT THE DIII D NATIONAL FUSION FACILITY

GA A23162 ENHANCED COMPUTATIONAL INFRASTRUCTURE FOR DATA ANALYSIS AT THE DIII D NATIONAL FUSION FACILITY GA A23162 ENHANCED COMPUTATIONAL INFRASTRUCTURE FOR DATA ANALYSIS AT THE DIII D NATIONAL FUSION FACILITY by D.P. SCHISSEL, Q. PENG, J. SCHACHTER, T.B. TEPSTRA T.A. CASPER, J. FREEMAN, R. JONG, K.M. KEITH,

More information

How To Create A Virtual Control Room Experiment

How To Create A Virtual Control Room Experiment Grids for Experimental Science: the Virtual Control Room K. Keahey 1,A, M. E. Papka 1,H, Q. Peng 2,I, D. Schissel 2,J G. Abla 2,B, T. Araki 1,3,C, J. Burruss 2,D, S. Feibush 4,G, P. Lane 1,E, S. Klasky,

More information

SiteCelerate white paper

SiteCelerate white paper SiteCelerate white paper Arahe Solutions SITECELERATE OVERVIEW As enterprises increases their investment in Web applications, Portal and websites and as usage of these applications increase, performance

More information

CHAPTER FIVE RESULT ANALYSIS

CHAPTER FIVE RESULT ANALYSIS CHAPTER FIVE RESULT ANALYSIS 5.1 Chapter Introduction 5.2 Discussion of Results 5.3 Performance Comparisons 5.4 Chapter Summary 61 5.1 Chapter Introduction This chapter outlines the results obtained from

More information

GT 6.0 GSI C Security: Key Concepts

GT 6.0 GSI C Security: Key Concepts GT 6.0 GSI C Security: Key Concepts GT 6.0 GSI C Security: Key Concepts Overview GSI uses public key cryptography (also known as asymmetric cryptography) as the basis for its functionality. Many of the

More information

Managing Credentials with

Managing Credentials with Managing Credentials with MyProxy Jim Basney National Center for Supercomputing Applications University of Illinois jbasney@ncsa.uiuc.edu http://myproxy.ncsa.uiuc.edu/ What is MyProxy? A service for managing

More information

Concepts and Architecture of Grid Computing. Advanced Topics Spring 2008 Prof. Robert van Engelen

Concepts and Architecture of Grid Computing. Advanced Topics Spring 2008 Prof. Robert van Engelen Concepts and Architecture of Grid Computing Advanced Topics Spring 2008 Prof. Robert van Engelen Overview Grid users: who are they? Concept of the Grid Challenges for the Grid Evolution of Grid systems

More information

MDSplus Automated Build and Distribution System

MDSplus Automated Build and Distribution System PSFC/JA-13-23 MDSplus Automated Build and Distribution System Fredian T.W., Stillerman J.A.*, Manduchi G.** * Plasma Science and Fusion Center, MIT ** Consorzio RFX, Euratom-ENEA Association, Padova,Italy

More information

Grid Scheduling Dictionary of Terms and Keywords

Grid Scheduling Dictionary of Terms and Keywords Grid Scheduling Dictionary Working Group M. Roehrig, Sandia National Laboratories W. Ziegler, Fraunhofer-Institute for Algorithms and Scientific Computing Document: Category: Informational June 2002 Status

More information

GridFTP: A Data Transfer Protocol for the Grid

GridFTP: A Data Transfer Protocol for the Grid GridFTP: A Data Transfer Protocol for the Grid Grid Forum Data Working Group on GridFTP Bill Allcock, Lee Liming, Steven Tuecke ANL Ann Chervenak USC/ISI Introduction In Grid environments,

More information

Authentication is not Authorization?! And what is a "digital signature" anyway?

Authentication is not Authorization?! And what is a digital signature anyway? Authentication is not Authorization?! And what is a "digital signature" anyway? Prepared by R. David Vernon Revised 12/01 Introduction REV 1A As part of the IT Architecture Initiative, the Office of Information

More information

Concepts and Architecture of the Grid. Summary of Grid 2, Chapter 4

Concepts and Architecture of the Grid. Summary of Grid 2, Chapter 4 Concepts and Architecture of the Grid Summary of Grid 2, Chapter 4 Concepts of Grid Mantra: Coordinated resource sharing and problem solving in dynamic, multi-institutional virtual organizations Allows

More information

Sync Security and Privacy Brief

Sync Security and Privacy Brief Introduction Security and privacy are two of the leading issues for users when transferring important files. Keeping data on-premises makes business and IT leaders feel more secure, but comes with technical

More information

The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets

The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets!! Large data collections appear in many scientific domains like climate studies.!! Users and

More information

Protect Microsoft Exchange databases, achieve long-term data retention

Protect Microsoft Exchange databases, achieve long-term data retention Technical white paper Protect Microsoft Exchange databases, achieve long-term data retention HP StoreOnce Backup systems, HP StoreOnce Catalyst, and Symantec NetBackup OpenStorage Table of contents Introduction...

More information

Cluster, Grid, Cloud Concepts

Cluster, Grid, Cloud Concepts Cluster, Grid, Cloud Concepts Kalaiselvan.K Contents Section 1: Cluster Section 2: Grid Section 3: Cloud Cluster An Overview Need for a Cluster Cluster categorizations A computer cluster is a group of

More information

MIGRATING DESKTOP AND ROAMING ACCESS. Migrating Desktop and Roaming Access Whitepaper

MIGRATING DESKTOP AND ROAMING ACCESS. Migrating Desktop and Roaming Access Whitepaper Migrating Desktop and Roaming Access Whitepaper Poznan Supercomputing and Networking Center Noskowskiego 12/14 61-704 Poznan, POLAND 2004, April white-paper-md-ras.doc 1/11 1 Product overview In this whitepaper

More information

ORACLE DATABASE 10G ENTERPRISE EDITION

ORACLE DATABASE 10G ENTERPRISE EDITION ORACLE DATABASE 10G ENTERPRISE EDITION OVERVIEW Oracle Database 10g Enterprise Edition is ideal for enterprises that ENTERPRISE EDITION For enterprises of any size For databases up to 8 Exabytes in size.

More information

Deploying a distributed data storage system on the UK National Grid Service using federated SRB

Deploying a distributed data storage system on the UK National Grid Service using federated SRB Deploying a distributed data storage system on the UK National Grid Service using federated SRB Manandhar A.S., Kleese K., Berrisford P., Brown G.D. CCLRC e-science Center Abstract As Grid enabled applications

More information

Data Management in an International Data Grid Project. Timur Chabuk 04/09/2007

Data Management in an International Data Grid Project. Timur Chabuk 04/09/2007 Data Management in an International Data Grid Project Timur Chabuk 04/09/2007 Intro LHC opened in 2005 several Petabytes of data per year data created at CERN distributed to Regional Centers all over the

More information

IBM Global Technology Services September 2007. NAS systems scale out to meet growing storage demand.

IBM Global Technology Services September 2007. NAS systems scale out to meet growing storage demand. IBM Global Technology Services September 2007 NAS systems scale out to meet Page 2 Contents 2 Introduction 2 Understanding the traditional NAS role 3 Gaining NAS benefits 4 NAS shortcomings in enterprise

More information

A New Standard in Encrypted Email. A discussion on push, pull and transparent delivery

A New Standard in Encrypted Email. A discussion on push, pull and transparent delivery A New Standard in Encrypted Email A discussion on push, pull and transparent delivery By ZixCorp November 2010 2 Email enhances our daily business life. It enables efficient, real-time communication, unites

More information

Web Service Robust GridFTP

Web Service Robust GridFTP Web Service Robust GridFTP Sang Lim, Geoffrey Fox, Shrideep Pallickara and Marlon Pierce Community Grid Labs, Indiana University 501 N. Morton St. Suite 224 Bloomington, IN 47404 {sblim, gcf, spallick,

More information

Making a Case for Including WAN Optimization in your Global SharePoint Deployment

Making a Case for Including WAN Optimization in your Global SharePoint Deployment Making a Case for Including WAN Optimization in your Global SharePoint Deployment Written by: Mauro Cardarelli Mauro Cardarelli is co-author of "Essential SharePoint 2007 -Delivering High Impact Collaboration"

More information

Accelerate > Converged Storage Infrastructure. DDN Case Study. ddn.com. 2013 DataDirect Networks. All Rights Reserved

Accelerate > Converged Storage Infrastructure. DDN Case Study. ddn.com. 2013 DataDirect Networks. All Rights Reserved DDN Case Study Accelerate > Converged Storage Infrastructure 2013 DataDirect Networks. All Rights Reserved The University of Florida s (ICBR) offers access to cutting-edge technologies designed to enable

More information

Grid Computing Vs. Cloud Computing

Grid Computing Vs. Cloud Computing International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 6 (2013), pp. 577-582 International Research Publications House http://www. irphouse.com /ijict.htm Grid

More information

An Oracle White Paper November 2010. Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics

An Oracle White Paper November 2010. Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics An Oracle White Paper November 2010 Leveraging Massively Parallel Processing in an Oracle Environment for Big Data Analytics 1 Introduction New applications such as web searches, recommendation engines,

More information

JAVA-BASED FRAMEWORK FOR REMOTE ACCESS TO LABORATORY EXPERIMENTS. Department of Electrical Engineering University of Hagen D-58084 Hagen, Germany

JAVA-BASED FRAMEWORK FOR REMOTE ACCESS TO LABORATORY EXPERIMENTS. Department of Electrical Engineering University of Hagen D-58084 Hagen, Germany JAVA-BASED FRAMEWORK FOR REMOTE ACCESS TO LABORATORY EXPERIMENTS Christof Röhrig, 1 Andreas Jochheim 2 Department of Electrical Engineering University of Hagen D-58084 Hagen, Germany Abstract: This paper

More information

Star System. 2004 Deitel & Associates, Inc. All rights reserved.

Star System. 2004 Deitel & Associates, Inc. All rights reserved. Star System Apple Macintosh 1984 First commercial OS GUI Chapter 1 Introduction to Operating Systems Outline 1.1 1.2 1.3 1.4 1.5 1.6 1.7 1.8 1.9 1.10 1.11 1.12 Introduction What Is an Operating System?

More information

GA A23745 STATUS OF THE LINUX PC CLUSTER FOR BETWEEN-PULSE DATA ANALYSES AT DIII D

GA A23745 STATUS OF THE LINUX PC CLUSTER FOR BETWEEN-PULSE DATA ANALYSES AT DIII D GA A23745 STATUS OF THE LINUX PC CLUSTER FOR BETWEEN-PULSE by Q. PENG, R.J. GROEBNER, L.L. LAO, J. SCHACHTER, D.P. SCHISSEL, and M.R. WADE AUGUST 2001 DISCLAIMER This report was prepared as an account

More information

Axceleon s CloudFuzion Turbocharges 3D Rendering On Amazon s EC2

Axceleon s CloudFuzion Turbocharges 3D Rendering On Amazon s EC2 Axceleon s CloudFuzion Turbocharges 3D Rendering On Amazon s EC2 In the movie making, visual effects and 3D animation industrues meeting project and timing deadlines is critical to success. Poor quality

More information

NETWORK ISSUES: COSTS & OPTIONS

NETWORK ISSUES: COSTS & OPTIONS VIDEO CONFERENCING NETWORK ISSUES: COSTS & OPTIONS Prepared By: S. Ann Earon, Ph.D., President Telemanagement Resources International Inc. Sponsored by Vidyo By:S.AnnEaron,Ph.D. Introduction Successful

More information

Installation Guide NetIQ AppManager

Installation Guide NetIQ AppManager Installation Guide NetIQ AppManager April 2016 www.netiq.com/documentation Legal Notice NetIQ AppManager is covered by United States Patent No(s): 05829001, 05986653, 05999178, 06078324, 06397359, 06408335.

More information

Building Remote Access VPNs

Building Remote Access VPNs Building Remote Access VPNs 124 Grove Street, Suite 309 Franklin, MA 02038 877-4-ALTIGA www.altiga.com Building Remote Access VPNs: Harnessing the Power of the Internet to Reduce Costs and Boost Performance

More information

Abstract. 1. Introduction. Ohio State University Columbus, OH 43210 {langella,oster,hastings,kurc,saltz}@bmi.osu.edu

Abstract. 1. Introduction. Ohio State University Columbus, OH 43210 {langella,oster,hastings,kurc,saltz}@bmi.osu.edu Dorian: Grid Service Infrastructure for Identity Management and Federation Stephen Langella 1, Scott Oster 1, Shannon Hastings 1, Frank Siebenlist 2, Tahsin Kurc 1, Joel Saltz 1 1 Department of Biomedical

More information

Four Ways High-Speed Data Transfer Can Transform Oil and Gas WHITE PAPER

Four Ways High-Speed Data Transfer Can Transform Oil and Gas WHITE PAPER Transform Oil and Gas WHITE PAPER TABLE OF CONTENTS Overview Four Ways to Accelerate the Acquisition of Remote Sensing Data Maximize HPC Utilization Simplify and Optimize Data Distribution Improve Business

More information

An approach to grid scheduling by using Condor-G Matchmaking mechanism

An approach to grid scheduling by using Condor-G Matchmaking mechanism An approach to grid scheduling by using Condor-G Matchmaking mechanism E. Imamagic, B. Radic, D. Dobrenic University Computing Centre, University of Zagreb, Croatia {emir.imamagic, branimir.radic, dobrisa.dobrenic}@srce.hr

More information

GA A22713 SOFTWARE DEVELOPMENT ON THE DIII D CONTROL AND DATA ACQUISITION COMPUTERS

GA A22713 SOFTWARE DEVELOPMENT ON THE DIII D CONTROL AND DATA ACQUISITION COMPUTERS GA A22713 SOFTWARE DEVELOPMENT ON THE DIII D CONTROL AND DATA ACQUISITION COMPUTERS by B.G. PENAFLOR, B.B. McHARG, JR., and D. PIGLOWSKI NOVEMBER 1997 DISCLAIMER This report was prepared as an account

More information

Science DMZs Understanding their role in high-performance data transfers

Science DMZs Understanding their role in high-performance data transfers Science DMZs Understanding their role in high-performance data transfers Chris Tracy, Network Engineer Eli Dart, Network Engineer ESnet Engineering Group Overview Bulk Data Movement a common task Pieces

More information

Transition Networks White Paper. Network Security. Why Authentication Matters YOUR NETWORK. OUR CONNECTION.

Transition Networks White Paper. Network Security. Why Authentication Matters YOUR NETWORK. OUR CONNECTION. Transition Networks White Paper Why Authentication Matters YOUR NETWORK. OUR CONNECTION. : Why Authentication Matters For most organizations physical security is a given. Whether it is video surveillance,

More information

DIGITAL RIGHTS MANAGEMENT SYSTEM FOR MULTIMEDIA FILES

DIGITAL RIGHTS MANAGEMENT SYSTEM FOR MULTIMEDIA FILES DIGITAL RIGHTS MANAGEMENT SYSTEM FOR MULTIMEDIA FILES Saiprasad Dhumal * Prof. K.K. Joshi Prof Sowmiya Raksha VJTI, Mumbai. VJTI, Mumbai VJTI, Mumbai. Abstract piracy of digital content is a one of the

More information

From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller

From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller White Paper From Ethernet Ubiquity to Ethernet Convergence: The Emergence of the Converged Network Interface Controller The focus of this paper is on the emergence of the converged network interface controller

More information

Entrust Managed Services PKI. Getting started with digital certificates and Entrust Managed Services PKI. Document issue: 1.0

Entrust Managed Services PKI. Getting started with digital certificates and Entrust Managed Services PKI. Document issue: 1.0 Entrust Managed Services PKI Getting started with digital certificates and Entrust Managed Services PKI Document issue: 1.0 Date of issue: May 2009 Copyright 2009 Entrust. All rights reserved. Entrust

More information

HPC & Visualization. Visualization and High-Performance Computing

HPC & Visualization. Visualization and High-Performance Computing HPC & Visualization Visualization and High-Performance Computing Visualization is a critical step in gaining in-depth insight into research problems, empowering understanding that is not possible with

More information

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago Globus Striped GridFTP Framework and Server Raj Kettimuthu, ANL and U. Chicago Outline Introduction Features Motivation Architecture Globus XIO Experimental Results 3 August 2005 The Ohio State University

More information

Big data management with IBM General Parallel File System

Big data management with IBM General Parallel File System Big data management with IBM General Parallel File System Optimize storage management and boost your return on investment Highlights Handles the explosive growth of structured and unstructured data Offers

More information

Online Transaction Processing in SQL Server 2008

Online Transaction Processing in SQL Server 2008 Online Transaction Processing in SQL Server 2008 White Paper Published: August 2007 Updated: July 2008 Summary: Microsoft SQL Server 2008 provides a database platform that is optimized for today s applications,

More information

Advanced Tools for Enhancing Control Room Collaborations

Advanced Tools for Enhancing Control Room Collaborations Advanced Tools for Enhancing Control Room Collaborations G. Abla a,, S.M. Flanagan a, Q. Peng a, J.R. Burruss a, D.P. Schissel a a General Atomics, P.O. Box 85608, San Diego, California 92186-5608 Abstract

More information

White paper: Unlocking the potential of load testing to maximise ROI and reduce risk.

White paper: Unlocking the potential of load testing to maximise ROI and reduce risk. White paper: Unlocking the potential of load testing to maximise ROI and reduce risk. Executive Summary Load testing can be used in a range of business scenarios to deliver numerous benefits. At its core,

More information

White Paper. Interactive Multicast Technology. Changing the Rules of Enterprise Streaming Video

White Paper. Interactive Multicast Technology. Changing the Rules of Enterprise Streaming Video Interactive Multicast Technology Changing the Rules of Enterprise Streaming Video V Table of Contents Introduction IP Multicast Technology in a Nutshell The Missing Interactivity: VCON Interactive Multicast

More information

White Paper. Enterprise IPTV and Video Streaming with the Blue Coat ProxySG >

White Paper. Enterprise IPTV and Video Streaming with the Blue Coat ProxySG > White Paper Enterprise IPTV and Video Streaming with the Blue Coat ProxySG > Table of Contents INTRODUCTION................................................... 2 SOLUTION ARCHITECTURE.........................................

More information

Chapter 7. Using Hadoop Cluster and MapReduce

Chapter 7. Using Hadoop Cluster and MapReduce Chapter 7 Using Hadoop Cluster and MapReduce Modeling and Prototyping of RMS for QoS Oriented Grid Page 152 7. Using Hadoop Cluster and MapReduce for Big Data Problems The size of the databases used in

More information

Scientific versus Business Workflows

Scientific versus Business Workflows 2 Scientific versus Business Workflows Roger Barga and Dennis Gannon The formal concept of a workflow has existed in the business world for a long time. An entire industry of tools and technology devoted

More information

Chapter 17. Transport-Level Security

Chapter 17. Transport-Level Security Chapter 17 Transport-Level Security Web Security Considerations The World Wide Web is fundamentally a client/server application running over the Internet and TCP/IP intranets The following characteristics

More information

Cisco Application Networking for Citrix Presentation Server

Cisco Application Networking for Citrix Presentation Server Cisco Application Networking for Citrix Presentation Server Faster Site Navigation, Less Bandwidth and Server Processing, and Greater Availability for Global Deployments What You Will Learn To address

More information

10 Secure Electronic Transactions: Overview, Capabilities, and Current Status

10 Secure Electronic Transactions: Overview, Capabilities, and Current Status 10 Secure Electronic Transactions: Overview, Capabilities, and Current Status Gordon Agnew A&F Consulting, and University of Waterloo, Ontario, Canada 10.1 Introduction Until recently, there were two primary

More information

THE CCLRC DATA PORTAL

THE CCLRC DATA PORTAL THE CCLRC DATA PORTAL Glen Drinkwater, Shoaib Sufi CCLRC Daresbury Laboratory, Daresbury, Warrington, Cheshire, WA4 4AD, UK. E-mail: g.j.drinkwater@dl.ac.uk, s.a.sufi@dl.ac.uk Abstract: The project aims

More information

Using DeployR to Solve the R Integration Problem

Using DeployR to Solve the R Integration Problem DEPLOYR WHITE PAPER Using DeployR to olve the R Integration Problem By the Revolution Analytics DeployR Team March 2015 Introduction Organizations use analytics to empower decision making, often in real

More information

Protecting enterprise servers with StoreOnce and CommVault Simpana

Protecting enterprise servers with StoreOnce and CommVault Simpana Technical white paper Protecting enterprise servers with StoreOnce and CommVault Simpana HP StoreOnce Backup systems Table of contents Introduction 2 Technology overview 2 HP StoreOnce Backup systems key

More information

An Integrated CyberSecurity Approach for HEP Grids. Workshop Report. http://hpcrd.lbl.gov/hepcybersecurity/

An Integrated CyberSecurity Approach for HEP Grids. Workshop Report. http://hpcrd.lbl.gov/hepcybersecurity/ An Integrated CyberSecurity Approach for HEP Grids Workshop Report http://hpcrd.lbl.gov/hepcybersecurity/ 1. Introduction The CMS and ATLAS experiments at the Large Hadron Collider (LHC) being built at

More information

Assignment # 1 (Cloud Computing Security)

Assignment # 1 (Cloud Computing Security) Assignment # 1 (Cloud Computing Security) Group Members: Abdullah Abid Zeeshan Qaiser M. Umar Hayat Table of Contents Windows Azure Introduction... 4 Windows Azure Services... 4 1. Compute... 4 a) Virtual

More information

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM Sneha D.Borkar 1, Prof.Chaitali S.Surtakar 2 Student of B.E., Information Technology, J.D.I.E.T, sborkar95@gmail.com Assistant Professor, Information

More information

White Paper Delivering Web Services Security: The Entrust Secure Transaction Platform

White Paper Delivering Web Services Security: The Entrust Secure Transaction Platform White Paper Delivering Web Services Security: September 2003 Copyright 2003 Entrust. All rights reserved. Entrust is a registered trademark of Entrust, Inc. in the United States and certain other countries.

More information

Data Management/Visualization on the Grid at PPPL. Scott A. Klasky Stephane Ethier Ravi Samtaney

Data Management/Visualization on the Grid at PPPL. Scott A. Klasky Stephane Ethier Ravi Samtaney Data Management/Visualization on the Grid at PPPL Scott A. Klasky Stephane Ethier Ravi Samtaney The Problem Simulations at NERSC generate GB s TB s of data. The transfer time for practical visualization

More information

Unicenter Remote Control r11

Unicenter Remote Control r11 Data Sheet Unicenter Remote Control r11 Unicenter Remote Control TM is a highly reliable and secure application for controlling and supporting remote Windows and Linux systems. It delivers all of the features

More information

The Lattice Project: A Multi-Model Grid Computing System. Center for Bioinformatics and Computational Biology University of Maryland

The Lattice Project: A Multi-Model Grid Computing System. Center for Bioinformatics and Computational Biology University of Maryland The Lattice Project: A Multi-Model Grid Computing System Center for Bioinformatics and Computational Biology University of Maryland Parallel Computing PARALLEL COMPUTING a form of computation in which

More information

IBM SAP International Competence Center. Load testing SAP ABAP Web Dynpro applications with IBM Rational Performance Tester

IBM SAP International Competence Center. Load testing SAP ABAP Web Dynpro applications with IBM Rational Performance Tester IBM SAP International Competence Center Load testing SAP ABAP Web Dynpro applications with IBM Rational Performance Tester Ease of use, excellent technical support from the IBM Rational team and, of course,

More information

Archive Data Retention & Compliance. Solutions Integrated Storage Appliances. Management Optimized Storage & Migration

Archive Data Retention & Compliance. Solutions Integrated Storage Appliances. Management Optimized Storage & Migration Solutions Integrated Storage Appliances Management Optimized Storage & Migration Archive Data Retention & Compliance Services Global Installation & Support SECURING THE FUTURE OF YOUR DATA w w w.q sta

More information

Background on Elastic Compute Cloud (EC2) AMI s to choose from including servers hosted on different Linux distros

Background on Elastic Compute Cloud (EC2) AMI s to choose from including servers hosted on different Linux distros David Moses January 2014 Paper on Cloud Computing I Background on Tools and Technologies in Amazon Web Services (AWS) In this paper I will highlight the technologies from the AWS cloud which enable you

More information

Overview. SSL Cryptography Overview CHAPTER 1

Overview. SSL Cryptography Overview CHAPTER 1 CHAPTER 1 Note The information in this chapter applies to both the ACE module and the ACE appliance unless otherwise noted. The features in this chapter apply to IPv4 and IPv6 unless otherwise noted. Secure

More information

Understanding the Benefits of IBM SPSS Statistics Server

Understanding the Benefits of IBM SPSS Statistics Server IBM SPSS Statistics Server Understanding the Benefits of IBM SPSS Statistics Server Contents: 1 Introduction 2 Performance 101: Understanding the drivers of better performance 3 Why performance is faster

More information

GoToMyPC Corporate Advanced Firewall Support Features

GoToMyPC Corporate Advanced Firewall Support Features F A C T S H E E T GoToMyPC Corporate Advanced Firewall Support Features Citrix GoToMyPC Corporate features Citrix Online s advanced connectivity technology. We support all of the common firewall and proxy

More information

EMC DOCUMENTUM MANAGING DISTRIBUTED ACCESS

EMC DOCUMENTUM MANAGING DISTRIBUTED ACCESS EMC DOCUMENTUM MANAGING DISTRIBUTED ACCESS This white paper describes the various distributed architectures supported by EMC Documentum and the relative merits and demerits of each model. It can be used

More information

How To Build A Cloud Computer

How To Build A Cloud Computer Introducing the Singlechip Cloud Computer Exploring the Future of Many-core Processors White Paper Intel Labs Jim Held Intel Fellow, Intel Labs Director, Tera-scale Computing Research Sean Koehl Technology

More information

IO Informatics The Sentient Suite

IO Informatics The Sentient Suite IO Informatics The Sentient Suite Our software, The Sentient Suite, allows a user to assemble, view, analyze and search very disparate information in a common environment. The disparate data can be numeric

More information

Defense Technical Information Center Compilation Part Notice

Defense Technical Information Center Compilation Part Notice UNCLASSIFIED Defense Technical Information Center Compilation Part Notice ADP012353 TITLE: Advanced 3D Visualization Web Technology and its Use in Military and Intelligence Applications DISTRIBUTION: Approved

More information

A Federated Model for Secure Web-Based Videoconferencing

A Federated Model for Secure Web-Based Videoconferencing A Federated Model for Secure Web-Based Videoconferencing Douglas C. Sicker, Ameet Kulkarni, Anand Chavali, and Mudassir Fajandar Interdisciplinary Telecommunications Dept. and Dept. of Computer Science

More information

PC-Duo Web Console Installation Guide

PC-Duo Web Console Installation Guide PC-Duo Web Console Installation Guide Release 12.1 August 2012 Vector Networks, Inc. 541 Tenth Street, Unit 123 Atlanta, GA 30318 (800) 330-5035 http://www.vector-networks.com Copyright 2012 Vector Networks

More information

Transparent Identification of Users

Transparent Identification of Users Transparent Identification of Users Websense Web Security Solutions v7.5, v7.6 Transparent Identification of Users 1996 2011, Websense, Inc. All rights reserved. 10240 Sorrento Valley Rd., San Diego, CA

More information

CA XOsoft Content Distribution v4

CA XOsoft Content Distribution v4 PRODUCT BRIEF: CA XOSOFT CONTENT DISTRIBUTION CA XOsoft Content Distribution v4 CA XOSOFT CONTENT DISTRIBUTION (FORMERLY CA XOSOFT WANSYNC CD) IS A SIMPLE, HIGHLY FLEXIBLE AND COST-EFFECTIVE CONTENT DELIVERY,

More information

IoT Security Platform

IoT Security Platform IoT Security Platform 2 Introduction Wars begin when the costs of attack are low, the benefits for a victor are high, and there is an inability to enforce law. The same is true in cyberwars. Today there

More information

Quantum StorNext. Product Brief: Distributed LAN Client

Quantum StorNext. Product Brief: Distributed LAN Client Quantum StorNext Product Brief: Distributed LAN Client NOTICE This product brief may contain proprietary information protected by copyright. Information in this product brief is subject to change without

More information

Data Grids. Lidan Wang April 5, 2007

Data Grids. Lidan Wang April 5, 2007 Data Grids Lidan Wang April 5, 2007 Outline Data-intensive applications Challenges in data access, integration and management in Grid setting Grid services for these data-intensive application Architectural

More information

Email Security Solutions

Email Security Solutions TECHNOLOGY REPORT Email Security Solutions 1 TECHNOLOGY REPORT SUPPLEMENT EMAIL SECURITY TECHNOLOGY REPORT IF YOUR EMAIL IS SO CRITICAL, CAN YOU BE SURE IT S REALLY REALLY PRIVATE? FIND THE FULL RESULTS

More information

W H I T E P A P E R. Deriving Intelligence from Large Data Using Hadoop and Applying Analytics. Abstract

W H I T E P A P E R. Deriving Intelligence from Large Data Using Hadoop and Applying Analytics. Abstract W H I T E P A P E R Deriving Intelligence from Large Data Using Hadoop and Applying Analytics Abstract This white paper is focused on discussing the challenges facing large scale data processing and the

More information

IBM Deep Computing Visualization Offering

IBM Deep Computing Visualization Offering P - 271 IBM Deep Computing Visualization Offering Parijat Sharma, Infrastructure Solution Architect, IBM India Pvt Ltd. email: parijatsharma@in.ibm.com Summary Deep Computing Visualization in Oil & Gas

More information

A Next-Generation Analytics Ecosystem for Big Data. Colin White, BI Research September 2012 Sponsored by ParAccel

A Next-Generation Analytics Ecosystem for Big Data. Colin White, BI Research September 2012 Sponsored by ParAccel A Next-Generation Analytics Ecosystem for Big Data Colin White, BI Research September 2012 Sponsored by ParAccel BIG DATA IS BIG NEWS The value of big data lies in the business analytics that can be generated

More information

NetIQ AppManager for Self Monitoring UNIX and Linux Servers (AMHealthUNIX) Management Guide

NetIQ AppManager for Self Monitoring UNIX and Linux Servers (AMHealthUNIX) Management Guide NetIQ AppManager for Self Monitoring UNIX and Linux Servers (AMHealthUNIX) Management Guide September 2014 Legal Notice THIS DOCUMENT AND THE SOFTWARE DESCRIBED IN THIS DOCUMENT ARE FURNISHED UNDER AND

More information

Wyse Device Manager TM

Wyse Device Manager TM Wyse Device Manager TM Secure, flexible and effective thin client management software All youneed toknow.here. An introduction to Wyse Device Manager TM Minimize your desktop support costs. Maximize desktop

More information

AppSense Environment Manager. Enterprise Design Guide

AppSense Environment Manager. Enterprise Design Guide Enterprise Design Guide Contents Introduction... 3 Document Purpose... 3 Basic Architecture... 3 Common Components and Terminology... 4 Best Practices... 5 Scalability Designs... 6 Management Server Scalability...

More information

expertcity DesktopStreaming Like being there. Technology and Security

expertcity DesktopStreaming Like being there. Technology and Security Table of Contents Introduction Tools Overview Architecture Firewall Configuration Security Conclusion Technology and Security Introduction The Internet is rapidly changing the way both consumers and corporations

More information

Cloud Computing and Advanced Relationship Analytics

Cloud Computing and Advanced Relationship Analytics Cloud Computing and Advanced Relationship Analytics Using Objectivity/DB to Discover the Relationships in your Data By Brian Clark Vice President, Product Management Objectivity, Inc. 408 992 7136 brian.clark@objectivity.com

More information