Hari Subramoni. Education: Employment: Research Interests: Projects:
|
|
|
- Amberly Harrington
- 9 years ago
- Views:
Transcription
1 Hari Subramoni Senior Research Associate, Dept. of Computer Science and Engineering The Ohio State University, Columbus, OH Tel: (614) , Fax: (614) , subramon Education: B. Tech. June 2004 University College of Engineering, Trivandrum Computer Science and Engineering Ph.D. August 2013 The Ohio State University, Columbus, Ohio Computer Science and Engineering Employment: Aug 13 Current Senior Research Associate The Ohio State University Apr 08 Jul 13 Graduate Research Assistant The Ohio State University Jun 11 Aug 11 Temporary Student Intern Lawrence Livermore National Laboratory Oct 09 Dec 09 Research Intern IBM TJ Watson Research Center Dec 06 Aug 07 Member Technical Staff Force10 Networks, India Aug 04 Nov 06 Software Engineer ISoftTech-Sasken Solutions, India Research Interests: Parallel computer architecture, network-based computing, exascale computing, network topology aware computing, QoS, power-aware LAN-WAN communication, fault tolerance, virtualization, high performance job startup and cloud computing. Projects: Involved in design, development, testing and distribution (involved in various major/minor releases - 1.2p1 through 2.1rc1) of MVAPICH2/MVAPICH2-X software stack - an open-source implementation of the MPI-3.1 specification over modern highspeed networks such as InfiniBand, 10/40 GigE, iwarp and RDMA over Converged Ethernet (RoCE). This software is being used by more than 2,275 organizations world-wide in 74 countries and is powering some of the top supercomputing centers in the world, including the 7th ranked Stampede, 11th ranked Pleiades and the 15th ranked Tsubame 2.5. As of Jan 15, more than 230,000 downloads have taken place from this project s site. This software is also being distributed by many InfiniBand, 10GigE/iWARP and RoCE vendors in their software distributions. MVAPICH2-X software package provides support for hybrid MPI+PGAS (UPC and OpenSHMEM) programming models with unified communication runtime for emerging exascale systems. Some of the specific projects worked on while at The Ohio State University are: Scalable network topology discovery mechanism for InifniBand networks; Network topology aware, high performance MPI libraries; High performance, fault-tolerant MPI over modern multi-rail systems; High performance, power aware, file transfer mechanisms over wide area networks; Scalable middleware for financial applications using emerging networking protocols like AMQP; QoS aware high performance MPI; High performance job startup for MPI; High-performance and scalable middleware for Big-Data over InfiniBand, iwarp and RoCE. Studied the impact network topology can have on the performance of HPC applications and investigated techniques to lower the impact through the design of network topology aware, high performance MPI libraries during the summer internship at Lawrence Livermore National Laboratory. Looked at the impact NUMA architectures can have on the communication performance of low latency, on-line trading systems during the internship at IBM T.J Watson Research Center. Select Awards and Recognitions: Outstanding Graduate Research Award - Department of Computer Science and Engineering, The Ohio State University, First Author of Best Paper/Best Student Paper Finalist - The International Conference for High Performance Computing, Networking, Storage and Analysis (SC 2012).
2 IEEE Student Travel Award - IEEE Cluster Selected as Student Volunteer - International Conference for High Performance Computing, Networking, Storage and Analysis (SC 2008). Proposal Writing Experience: Extended help in various phases of proposal writing for the following successful proposals. Interacted with PIs from multiple institutions in defining overall proposal direction. Innovations to Transition a Campus Core Cyberinfrastructure to Serve Diverse and Emerging Researcher Needs, National Science Foundation (OCI), Unified Runtime for Supporting Hybrid Programming Models on Heterogeneous Architecture, National Science Foundation, SHF, A Comprehensive Performance Tuning Framework for the MPI Stack, National Science Foundation, SI2-SSI, HPC Application Energy Measurement and Optimization Tools, DOE SBIR/STTR Phase-I, Designing QoS-Aware MPI and File Systems Protocols for Emerging InfiniBand Clusters, National Science Foundation, Topology-Aware MPI Communication and Scheduling for Petascale Systems, National Science Foundation (STCI), Six more proposals (five to NSF and one to DOE) are under review. Publications: Invited Papers, Book Chapters, Journal Publications: 1. S. Sur, S. Potluri, K. Kandalla, H. Subramoni, K. Tomko, D. K. Panda, Co-Designing MPI Library and Applications for InfiniBand Clusters, Computer, 06 Sep IEEE Computer Society Digital Library. 2. D. K. Panda, S. Sur, H. Subramoni and K. Kandalla, Network Support for Collective Communication, Encyclopedia of Parallel Computing, Sep H. Subramoni, F. Petrini, V. Agarwal and D. Pasetto, Intra-Socket and Inter-Socket Communication in Multi-core Systems, Computer Architecture Letters 9(1): 13-16, Refereed Conference/Workshop Publications: 1. S. Chakraborty, H. Subramoni, A. Moody, A. Venkatesh, J. Perkins and D. K. Panda, Non-blocking PMI Extensions for Fast MPI Startup, Int l Symposium on Cluster, Cloud, and Grid Computing (CCGrid 2015), May 2015 Accepted to be published. 2. A. Venkatesh, H. Subramoni, K. Hamidouche and D. K. Panda, A High Performance Broadcast Design with Hardware Multicast and GPUDirect RDMA for Streaming Applications on InfiniBand Clusters, IEEE International Conference on High Performance Computing (HiPC 2014). Dec H. Subramoni, K. Kandalla, J. Jose, K. Tomko, K. Schulz, and D. Pekurovsky, Designing Topology Aware Communication Schedules for Alltoall Operations in Large InfiniBand Clusters, Int l Conference on Parallel Processing (ICPP 14). Sep S. Chakraborty, H. Subramoni, J. Perkins, A. Moody, M. Arnold, D. K. Panda, PMI Extensions for Scalable MPI Startup, EUROMPI 14. Sep H. Subramoni, K. Hamidouche, A. Venkatesh, S. Chakraborty and D. K. Panda, Designing MPI Library with Dynamic Connected Transport (DCT) of InfiniBand : Early Experiences. IEEE International Supercomputing Conference (ISC 14). June H. Subramoni, D. Bureddy, K. Kandalla, K. Schulz, B. Barth, J. Perkins, M. Arnold, and D. K. Panda, Design of Network Topology Aware Scheduling Services for Large InfiniBand Clusters, IEEE Cluster 13, Sep
3 7. S. Potluri, D. Bureddy, K. Hamidouche, A. Venkatesh, K. Kandalla, H. Subramoni, and D. K. Panda, MVAPICH- PRISM: A Proxy-based Communication Framework using InfiniBand and SCIF for Intel MIC Clusters, Int l Conference on Supercomputing (SC 13), Nov K. Hamidouche, S. Potluri, H. Subramoni, K. Kandalla and D. K. Panda, MIC-RO: Enabling Efficient Remote Offload on Heterogeneous Many Integrated Core (MIC) Clusters with InfiniBand, Int l Conference on Supercomputing (ICS 13), June S. Potluri, D. Bureddy, H. Wang, H. Subramoni and D. K. Panda, Extending OpenSHMEM for GPU Computing, Int l Parallel and Distributed Processing Symposium (IPDPS 13), May H. Subramoni, S. Potluri, K. Kandalla, B. Barth, J. Vienne, J. Keasler, K. Tomko, K. Schulz, A. Moody and D. K. Panda, Design of a Scalable InfiniBand Topology Service to Enable Network-Topology-Aware Placement of Processes, Int l Conference on Supercomputing (SC 12), Nov (Best Paper and Best Student Paper Finalist) 11. H. Subramoni, J. Vienne and D. K. Panda, A Scalable InfiniBand Network-Topology-Aware Performance Analysis Tool for MPI, Int l Workshop on Productivity and Performance (Proper 12), in conjunction with EuroPar, Aug N. S. Islam, M. W. Rahman, J. Jose, R. Rajachandrasekar, H. Wang, H. Subramoni, C. Murthy and D. K. Panda, High Performance RDMA-Based Design of HDFS over InfiniBand, Int l Conference on Supercomputing (SC 12), Nov R. Rajachandrasekar, J. Jaswani, H. Subramoni and D. K. Panda, Minimizing Network Contention in InfiniBand Clusters with a QoS-Aware Data-Staging Framework, Cluster (Cluster 12), Sep K. Kandalla, A. Buluç, H. Subramoni, K. Tomko, J. Vienne, L. Oliker, and D. K. Panda, Can Network-Offload based Non-Blocking Neighborhood MPI Collectives Improve Communication Overheads of Irregular Graph Algorithms?. International Workshop on Parallel Algorithms and Parallel Software (IWPAPS 2012), in conjunction with Cluster, Sep J. Huang, X. Ouyang, J. Jose, M. W. Rahman, H. Wang, M. Luo, H. Subramoni, C. Murthy and D. K. Panda, High- Performance Design of HBase with RDMA over InfiniBand, Int l Parallel and Distributed Processing Symposium (IPDPS 12), May K. Kandalla, U. Yang, J. Keasler, T. Kolev, A. Moody, H. Subramoni, K. Tomko, J. Vienne and D. K. Panda, Designing Non-blocking Allreduce with Collective Offload on InfiniBand Clusters: A Case Study with Conjugate Gradient Solvers Int l Parallel and Distributed Processing Symposium (IPDPS 12), May S. P. Raikar, H. Subramoni, K. Kandalla, J. Vienne and D. K. Panda, Designing Network Failover and Recovery in MPI for Multi-Rail InfiniBand Clusters, Int l Workshop on System Management Techniques, Processes, and Services (SMTPS), in conjunction with Int l Parallel and Distributed Processing Symposium (IPDPS 12), May J. Jose, H. Subramoni, K. Kandalla, M. W. Rahman, H. Wang, S. Narravula and D. K. Panda, Scalable Memcached design for InfiniBand Clusters using Hybrid Transports, Int l Symposium on Cluster, Cloud, and Grid Computing (CCGrid 2012), May H. Subramoni, K. Kandalla, J. Vienne, S. Sur, B. Barth, K. Tomko, R. McLay, K. Schulz and D. K. Panda, Design and Evaluation of Network Topology-/Speed-Aware Broadcast Algorithms for InfiniBand Clusters, Cluster 11, Sep J. Jose, H. Subramoni, M. Luo, M. Zhang, J. Huang, M. W. Rahman, N. S. Islam, X. Ouyang, H. Wang, S. Sur and D. K. Panda, Memcached Design on High Performance RDMA Capable Interconnects, Int l Conference on Parallel Processing (ICPP 11), Sep N. Dandapanthula, H. Subramoni, J. Vienne, K. Kandalla, S. Sur, D. K. Panda, and R. Brightwell, INAM - A Scalable InfiniBand Network Analysis and Monitoring Tool, 4th Int l Workshop on Productivity and Performance (PROPER 2011), in conjunction with EuroPar, Aug K. Kandalla, H. Subramoni, J. Vienne, K. Tomko, S. Sur and D. K. Panda, Designing Non-blocking Broadcast with Collective Offload on InfiniBand Clusters: A Case Study with HPL, Hot Interconnect 11, Aug K. Kandalla, H. Subramoni,, K. Tomko, D. Pekurovsky, S. Sur and D. K. Panda, High-Performance and Scalable Non- Blocking All-to-All with Collective Offload on InfiniBand Clusters: A Study with Parallel 3D FFT, Int l Supercomputing Conference (ISC), June 2011.
4 24. H. Subramoni, P. Lai, S. Sur and D. K. Panda, Improving Application Performance and Predictability using Multiple Virtual Lanes in Modern Multi-Core InfiniBand Clusters, Int l Conference on Parallel Processing (ICPP 10), Sep M. Luo, S. Potluri, P. Lai, E. P. Mancini, H. Subramoni, K. Kandalla, S. Sur and D. K. Panda, High Performance Design and Implementation of Nemesis Communication Layer for Two-sided and One-Sided MPI, The Workshop on Parallel Programming Models and System Software (P2S2 2012), in conjunction with Int l Conference on Parallel Processing (ICPP 10), Sep H. Subramoni, K. Kandalla, S. Sur and D. K. Panda, Design and Evaluation of Generalized Collective Communication Primitives with Overlap using ConnectX-2 Offload Engine, Int l Symposium on Hot Interconnects (HotI), Aug H. Subramoni, P. Lai, R. Kettimuthu and D. K. Panda, High Performance Data Transfer in Grid Environment Using GridFTP over InfiniBand, Int l Symposium on Cluster Computing and the Grid (CCGrid 10), May H. Subramoni, F. Petrini, V. Agarwal and D. Pasetto, Streaming, low-latency communication in on-line trading systems, Int l Workshop on System Management Techniques, Processes, and Services (SMTPS), in conjunction with Int l Parallel and Distributed Processing Symposium (IPDPS 10), April K. Kandalla, H. Subramoni, A. Vishnu and D. K. Panda, Designing Topology-Aware Collective Communication Algorithms for Large Scale InfiniBand Clusters: Case Studies with Scatter and Gather, The 10th Workshop on Communication Architecture for Clusters (CAC 10), in conjunction with Int l Parallel and Distributed Processing Symposium (IPDPS 10), April P. Lai, H. Subramoni, S. Narravula, A. Mamidala and D. K. Panda, Designing Efficient FTP Mechanisms for High Performance Data-Transfer over InfiniBand, Int l Conference on Parallel Processing (ICPP 09), Sep H. Subramoni, P. Lai, M. Luo, Dhabaleswar K. Panda, RDMA over Ethernet - A Preliminary Study, Workshop on High Performance Interconnects for Distributed Computing (HPIDC 09), in conjunction with Cluster, Sep H. Subramoni, M. Koop, and D. K. Panda, Designing Next Generation Clusters: Evaluation of InfiniBand DDR/QDR on Intel Computing Platforms, 17th Annual Symposium on High-Performance Interconnects (HotI 09), Aug K. Kandalla, H. Subramoni, G. Santhanaraman, M. Koop and D. K. Panda, The 9th Workshop on Communication Architecture for Clusters (CAC 09), in conjunction with Int l Parallel and Distributed Processing Symposium (IPDPS 09), May H. Subramoni, G. Marsh, S. Narravula, P. Lai and D.K. Panda, Design and Evaluation of Benchmarks for Financial Applications using Advanced Message Queuing Protocol (AMQP) over InfiniBand, Workshop on High Performance Computational Finance (in conjunction with SC 08), Austin, TX, November S. Narravula, H. Subramoni, P. Lai, R. Noronha and D. K. Panda Performance of HPC Middleware over InfiniBand WAN, Int l Conference on Parallel Processing (ICPP 08), Portland, Oregon, Sep Presentations: 1. H. Subramoni, K. Kandalla, J. Jose, K. Tomko, K. Schulz, and D. Pekurovsky, Designing Topology Aware Communication Schedules for Alltoall Operations in Large InfiniBand Clusters, Int l Conference on Parallel Processing (ICPP 14). Sep H. Subramoni, K. Hamidouche, A. Venkatesh, S. Chakraborty and D. K. Panda, Designing MPI Library with Dynamic Connected Transport (DCT) of InfiniBand : Early Experiences. IEEE International Supercomputing Conference (ISC 14). June H. Subramoni, D. Bureddy, K. Kandalla, K. Schulz, B. Barth, J. Perkins, M. Arnold, and D. K. Panda, Design of Network Topology Aware Scheduling Services for Large InfiniBand Clusters, IEEE Cluster 13, Sep H. Subramoni, S. Potluri, K. Kandalla, B. Barth, J. Vienne, J. Keasler, K. Tomko, K. Schulz, A. Moody, and D. K. Panda, Design of a Scalable InfiniBand Topology Service to Enable Network-Topology-Aware Placement of Processes, SC 12, November (Best Paper and Best Student Paper Finalist). 5. H. Subramoni, J. Vienne, and D. K. Panda, A Scalable InfiniBand Network Topology-Aware Performance Analysis Tool for MPI, 5th Workshop on Productivity and Performance (PROPER) Aug., H. Subramoni, K. Kandalla, J. Vienne, S. Sur, B. Barth, K. Tomko, R. McLay, K. Schulz and D. K. Panda, Design and Evaluation of Network Topology-/Speed-Aware Broadcast Algorithms for InfiniBand Clusters, Cluster 11, Sep
5 7. H. Subramoni, P. Lai, S. Sur and D. K. Panda, Improving Application Performance and Predictability using Multiple Virtual Lanes in Modern Multi-Core InfiniBand Clusters, Int l Conference on Parallel Processing (ICPP 10), Sep K. Kandalla, E. P. Mancini, S. Sur and D. K. Panda, Designing Power-Aware Collective Communication Algorithms for InfiniBand Clusters, Int l Conference on Parallel Processing (ICPP 10), Sep H. Subramoni, P. Lai, R. Kettimuthu and D. K. Panda, High Performance Data Transfer in Grid Environment Using GridFTP over InfiniBand, Int l Symposium on Cluster Computing and the Grid (CCGrid), May H. Subramoni, P. Lai, M. Luo, Dhabaleswar K. Panda, RDMA over Ethernet - A Preliminary Study, Workshop on High Performance Interconnects for Distributed Computing (HPIDC 09), in conjunction with Cluster, Sep H. Subramoni, M. Koop and D. K. Panda, Designing Next Generation Balanced Clusters with InfiniBand QDR and Intel Nehalem Architecture, Symposium on High Performance Interconnects (HOTI 09), Aug H. Subramoni, G. Marsh, S. Narravula, P. Lai and D. K. Panda, Design and Evaluation of Benchmarks for Financial Applications using Advanced Message Queueing Protocol (AMQP) over InfiniBand, Workshop on High Performance Computational Finance (WHPCF 08), in conjunction with SC 08, November Invited Tutorials: 1. InfiniBand and High-Speed Ethernet for Dummies, SuperComputing (SC 14), Nov Designing and Using High-End Computing Systems with InfiniBand and High-Speed Ethernet SuperComputing (SC 14), Nov Optimization and Tuning of MPI and PGAS Applications using MVAPICH2 and MVAPICH2-X XSEDE 14, July, InfiniBand and High-Speed Ethernet: Overview, Latest Status and Trends Int l Conference on High Performance Switching and Routing (HPSR 14), July, InfiniBand and High-Speed Ethernet: Overview, Latest Status and Trends Int l Supercomputing Conference (ISC 14), June, InfiniBand and High-Speed Ethernet for Dummies, SuperComputing (SC 13), Nov Advanced Topics in InfiniBand and High-speed Ethernet for Designing HEC Systems SuperComputing (SC 13), Nov InfiniBand and High-Speed Ethernet for Dummies, Int l Supercomputing Conference (ISC 13), June, Advanced Topics in InfiniBand and High-speed Ethernet for Designing HEC Systems Int l Supercomputing Conference (ISC 13), June, InfiniBand and High-Speed Ethernet for Dummies, SuperComputing (SC 12), Nov Designing High-End Computing Systems with InfiniBand and High-Speed Ethernet, SuperComputing (SC 12), Nov InfiniBand and High-Speed Ethernet for Dummies, Int l Supercomputing Conference (ISC 12), June, Designing High-End Computing Systems and Programming Models with InfiniBand and High-Speed Ethernet, Int l Supercomputing Conference (ISC 12), June, InfiniBand and High-Speed Ethernet for Dummies, SuperComputing (SC 11), Nov Designing High-End Computing Systems with InfiniBand and High-Speed Ethernet, SuperComputing (SC 11), Nov Invited Talks: 1. Optimizing and Tuning Techniques for Running MVAPICH2 over IB, InfiniBand User Group Meeting (IBUG), April MVAPICH2 Project Update and GPUDirect RDMA, HPC Advisory Council European Conference, June Experiencing HPC for Undergraduates - Graduate Student Perspective, SuperComputing (SC 12), Nov MVAPICH2 Project Update and Big Data Acceleration, HPC Advisory Council European Conference, June 2012.
6 Selected Professional Activities: IEEE Member Session Chair - Int l Workshop on Parallel Programming Models and Systems Software for High-End Computing (P2S2), 2014 Program Committee Member - International Conference on Parallel Processing (ICPP), 2014 Organizing Committee Member - MVAPICH User Group Meeting (MUG), 2013, 2014 Attendee - MPI Forum (MPI-3.1) Reviewer, SuperComputing (SC), 2013 Reviewer, Transactions on Parallel and Distributed Systems (TPDS), 2013 Reviewer, Journal of Parallel and Distributed Computing (JPDC), 2009, 2010, 2011, 2012, 2013, 2014 Reviewer, International Conference on Parallel Processing (ICPP), 2009, 2010, 2011 Reviewer, International Parallel & Distributed Processing Symposium (IPDPS), 2010, 2011, 2013 Reviewer, IEEE Cluster, 2010, 2011 Member of Career Guidance and Placement Cell at Under Graduate University
RDMA over Ethernet - A Preliminary Study
RDMA over Ethernet - A Preliminary Study Hari Subramoni, Miao Luo, Ping Lai and Dhabaleswar. K. Panda Computer Science & Engineering Department The Ohio State University Outline Introduction Problem Statement
Big Data: Hadoop and Memcached
Big Data: Hadoop and Memcached Talk at HPC Advisory Council Stanford Conference and Exascale Workshop (Feb 214) by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: [email protected] http://www.cse.ohio-state.edu/~panda
Can High-Performance Interconnects Benefit Memcached and Hadoop?
Can High-Performance Interconnects Benefit Memcached and Hadoop? D. K. Panda and Sayantan Sur Network-Based Computing Laboratory Department of Computer Science and Engineering The Ohio State University,
A Micro-benchmark Suite for Evaluating Hadoop RPC on High-Performance Networks
A Micro-benchmark Suite for Evaluating Hadoop RPC on High-Performance Networks Xiaoyi Lu, Md. Wasi- ur- Rahman, Nusrat Islam, and Dhabaleswar K. (DK) Panda Network- Based Compu2ng Laboratory Department
Accelerating Big Data Processing with Hadoop, Spark and Memcached
Accelerating Big Data Processing with Hadoop, Spark and Memcached Talk at HPC Advisory Council Stanford Conference (Feb 15) by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: [email protected]
Advancing Applications Performance With InfiniBand
Advancing Applications Performance With InfiniBand Pak Lui, Application Performance Manager September 12, 2013 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server and
Min Si. Argonne National Laboratory Mathematics and Computer Science Division
Min Si Contact Information Address 9700 South Cass Avenue, Bldg. 240, Lemont, IL 60439, USA Office +1 630-252-4249 Mobile +1 630-880-4388 E-mail [email protected] Homepage http://www.mcs.anl.gov/~minsi/ Current
Accelerating Spark with RDMA for Big Data Processing: Early Experiences
Accelerating Spark with RDMA for Big Data Processing: Early Experiences Xiaoyi Lu, Md. Wasi- ur- Rahman, Nusrat Islam, Dip7 Shankar, and Dhabaleswar K. (DK) Panda Network- Based Compu2ng Laboratory Department
High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand
High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand Hari Subramoni *, Ping Lai *, Raj Kettimuthu **, Dhabaleswar. K. (DK) Panda * * Computer Science and Engineering Department
How To Monitor Infiniband Network Data From A Network On A Leaf Switch (Wired) On A Microsoft Powerbook (Wired Or Microsoft) On An Ipa (Wired/Wired) Or Ipa V2 (Wired V2)
INFINIBAND NETWORK ANALYSIS AND MONITORING USING OPENSM N. Dandapanthula 1, H. Subramoni 1, J. Vienne 1, K. Kandalla 1, S. Sur 1, D. K. Panda 1, and R. Brightwell 2 Presented By Xavier Besseron 1 Date:
Accelerating Big Data Processing with Hadoop, Spark and Memcached
Accelerating Big Data Processing with Hadoop, Spark and Memcached Talk at HPC Advisory Council Switzerland Conference (Mar 15) by Dhabaleswar K. (DK) Panda The Ohio State University E-mail: [email protected]
Application and Micro-benchmark Performance using MVAPICH2-X on SDSC Gordon Cluster
Application and Micro-benchmark Performance using MVAPICH2-X on SDSC Gordon Cluster Mahidhar Tatineni ([email protected]) MVAPICH User Group Meeting August 27, 2014 NSF grants: OCI #0910847 Gordon: A Data
Performance Evaluation of InfiniBand with PCI Express
Performance Evaluation of InfiniBand with PCI Express Jiuxing Liu Server Technology Group IBM T. J. Watson Research Center Yorktown Heights, NY 1598 [email protected] Amith Mamidala, Abhinav Vishnu, and Dhabaleswar
Low-Latency Software Defined Network for High Performance Clouds
Low-Latency Software Defined Network for High Performance Clouds Paul Rad 1, Rajendra V. Boppana 2, Palden Lama 2, Gilad Berman 3 and Mo Jamshidi 1 1 Department of Computer Engineering, University of Texas
INAM - A Scalable InfiniBand Network Analysis and Monitoring Tool
INAM - A Scalable InfiniBand Network Analysis and Monitoring Tool N. Dandapanthula 1, H. Subramoni 1, J. Vienne 1, K. Kandalla 1, S. Sur 1, D. K. Panda 1, and R. Brightwell 2 1 Department of Computer Science
Designing Efficient FTP Mechanisms for High Performance Data-Transfer over InfiniBand
Designing Efficient FTP Mechanisms for High Performance Data-Transfer over InfiniBand Ping Lai, Hari Subramoni, Sundeep Narravula, Amith Mamidala, Dhabaleswar K. Panda Department of Computer Science and
Accelera'ng Big Data Processing with Hadoop, Spark and Memcached
Accelera'ng Big Data Processing with Hadoop, Spark and Memcached Talk at HPC Advisory Council Switzerland Conference (Mar 15) by Dhabaleswar K. (DK) Panda The Ohio State University E- mail: [email protected]
A Flexible Cluster Infrastructure for Systems Research and Software Development
Award Number: CNS-551555 Title: CRI: Acquisition of an InfiniBand Cluster with SMP Nodes Institution: Florida State University PIs: Xin Yuan, Robert van Engelen, Kartik Gopalan A Flexible Cluster Infrastructure
LS DYNA Performance Benchmarks and Profiling. January 2009
LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The
Why Compromise? A discussion on RDMA versus Send/Receive and the difference between interconnect and application semantics
Why Compromise? A discussion on RDMA versus Send/Receive and the difference between interconnect and application semantics Mellanox Technologies Inc. 2900 Stender Way, Santa Clara, CA 95054 Tel: 408-970-3400
Enabling High performance Big Data platform with RDMA
Enabling High performance Big Data platform with RDMA Tong Liu HPC Advisory Council Oct 7 th, 2014 Shortcomings of Hadoop Administration tooling Performance Reliability SQL support Backup and recovery
Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi
Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi ICPP 6 th International Workshop on Parallel Programming Models and Systems Software for High-End Computing October 1, 2013 Lyon, France
Enabling Multi-pipeline Data Transfer in HDFS for Big Data Applications
Enabling Multi-pipeline Data Transfer in HDFS for Big Data Applications Liqiang (Eric) Wang, Hong Zhang University of Wyoming Hai Huang IBM T.J. Watson Research Center Background Hadoop: Apache Hadoop
Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure. Abstract:
Performance Evaluation of the RDMA over Ethernet (RoCE) Standard in Enterprise Data Centers Infrastructure Motti Beck Director, Marketing [email protected] Michael Kagan Chief Technology Officer [email protected]
Mellanox Academy Online Training (E-learning)
Mellanox Academy Online Training (E-learning) 2013-2014 30 P age Mellanox offers a variety of training methods and learning solutions for instructor-led training classes and remote online learning (e-learning),
HPC enabling of OpenFOAM R for CFD applications
HPC enabling of OpenFOAM R for CFD applications Towards the exascale: OpenFOAM perspective Ivan Spisso 25-27 March 2015, Casalecchio di Reno, BOLOGNA. SuperComputing Applications and Innovation Department,
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
State of the Art Cloud Infrastructure
State of the Art Cloud Infrastructure Motti Beck, Director Enterprise Market Development WHD Global I April 2014 Next Generation Data Centers Require Fast, Smart Interconnect Software Defined Networks
High Throughput File Servers with SMB Direct, Using the 3 Flavors of RDMA network adapters
High Throughput File Servers with SMB Direct, Using the 3 Flavors of network adapters Jose Barreto Principal Program Manager Microsoft Corporation Abstract In Windows Server 2012, we introduce the SMB
Interconnecting Future DoE leadership systems
Interconnecting Future DoE leadership systems Rich Graham HPC Advisory Council, Stanford, 2015 HPC The Challenges 2 Proud to Accelerate Future DOE Leadership Systems ( CORAL ) Summit System Sierra System
Pedraforca: ARM + GPU prototype
www.bsc.es Pedraforca: ARM + GPU prototype Filippo Mantovani Workshop on exascale and PRACE prototypes Barcelona, 20 May 2014 Overview Goals: Test the performance, scalability, and energy efficiency of
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014
Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet Anand Rangaswamy September 2014 Storage Developer Conference Mellanox Overview Ticker: MLNX Leading provider of high-throughput,
SMB Direct for SQL Server and Private Cloud
SMB Direct for SQL Server and Private Cloud Increased Performance, Higher Scalability and Extreme Resiliency June, 2014 Mellanox Overview Ticker: MLNX Leading provider of high-throughput, low-latency server
Accelerating From Cluster to Cloud: Overview of RDMA on Windows HPC. Wenhao Wu Program Manager Windows HPC team
Accelerating From Cluster to Cloud: Overview of RDMA on Windows HPC Wenhao Wu Program Manager Windows HPC team Agenda Microsoft s Commitments to HPC RDMA for HPC Server RDMA for Storage in Windows 8 Microsoft
Solving I/O Bottlenecks to Enable Superior Cloud Efficiency
WHITE PAPER Solving I/O Bottlenecks to Enable Superior Cloud Efficiency Overview...1 Mellanox I/O Virtualization Features and Benefits...2 Summary...6 Overview We already have 8 or even 16 cores on one
Interoperability Testing and iwarp Performance. Whitepaper
Interoperability Testing and iwarp Performance Whitepaper Interoperability Testing and iwarp Performance Introduction In tests conducted at the Chelsio facility, results demonstrate successful interoperability
GASPI A PGAS API for Scalable and Fault Tolerant Computing
GASPI A PGAS API for Scalable and Fault Tolerant Computing Specification of a general purpose API for one-sided and asynchronous communication and provision of libraries, tools, examples and best practices
Introduction to Infiniband. Hussein N. Harake, Performance U! Winter School
Introduction to Infiniband Hussein N. Harake, Performance U! Winter School Agenda Definition of Infiniband Features Hardware Facts Layers OFED Stack OpenSM Tools and Utilities Topologies Infiniband Roadmap
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
Hadoop on the Gordon Data Intensive Cluster
Hadoop on the Gordon Data Intensive Cluster Amit Majumdar, Scientific Computing Applications Mahidhar Tatineni, HPC User Services San Diego Supercomputer Center University of California San Diego Dec 18,
SMB Advanced Networking for Fault Tolerance and Performance. Jose Barreto Principal Program Managers Microsoft Corporation
SMB Advanced Networking for Fault Tolerance and Performance Jose Barreto Principal Program Managers Microsoft Corporation Agenda SMB Remote File Storage for Server Apps SMB Direct (SMB over RDMA) SMB Multichannel
Cluster Implementation and Management; Scheduling
Cluster Implementation and Management; Scheduling CPS343 Parallel and High Performance Computing Spring 2013 CPS343 (Parallel and HPC) Cluster Implementation and Management; Scheduling Spring 2013 1 /
HPC Programming Framework Research Team
HPC Programming Framework Research Team 1. Team Members Naoya Maruyama (Team Leader) Motohiko Matsuda (Research Scientist) Soichiro Suzuki (Technical Staff) Mohamed Wahib (Postdoctoral Researcher) Shinichiro
Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks
WHITE PAPER July 2014 Achieving Real-Time Business Solutions Using Graph Database Technology and High Performance Networks Contents Executive Summary...2 Background...3 InfiniteGraph...3 High Performance
AT&T Global Network Client for Windows Product Support Matrix January 29, 2015
AT&T Global Network Client for Windows Product Support Matrix January 29, 2015 Product Support Matrix Following is the Product Support Matrix for the AT&T Global Network Client. See the AT&T Global Network
A SOFTWARE DEFINED NETWORKING ARCHITECTURE FOR HIGH PERFORMANCE CLOUDS 1
International Journal of Complex Systems Computing, Sensing and Control Vol. X, No. X, pp. 1-X, 201X Copyright 2015, TSI Press Printed in the USA. All rights reserved A SOFTWARE DEFINED NETWORKING ARCHITECTURE
RoCE vs. iwarp Competitive Analysis
WHITE PAPER August 21 RoCE vs. iwarp Competitive Analysis Executive Summary...1 RoCE s Advantages over iwarp...1 Performance and Benchmark Examples...3 Best Performance for Virtualization...4 Summary...
FTP 100: An Ultra-High Speed Data Transfer Service over Next Generation 100 Gigabit Per Second Network
FTP 100: An Ultra-High Speed Data Transfer Service over Next Generation 100 Gigabit Per Second Network Award Number: 53662 - US Department o f Energy Aw ard Title: ARRA: TAS::89 0227:: TASRecovery A ct
Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 SMB Direct
Mellanox Cloud and Database Acceleration Solution over Windows Server 2012 Direct Increased Performance, Scaling and Resiliency July 2012 Motti Beck, Director, Enterprise Market Development [email protected]
Deep Learning GPU-Based Hardware Platform
Deep Learning GPU-Based Hardware Platform Hardware and Software Criteria and Selection Mourad Bouache Yahoo! Performance Engineering Group Sunnyvale, CA +1.408.784.1446 [email protected] John Glover
3rd International Symposium on Big Data and Cloud Computing Challenges (ISBCC-2016) March 10-11, 2016 VIT University, Chennai, India
3rd International Symposium on Big Data and Cloud Computing Challenges (ISBCC-2016) March 10-11, 2016 VIT University, Chennai, India Call for Papers Cloud computing has emerged as a de facto computing
Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA
WHITE PAPER April 2014 Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA Executive Summary...1 Background...2 File Systems Architecture...2 Network Architecture...3 IBM BigInsights...5
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control EP/K006487/1 UK PI: Prof Gareth Taylor (BU) China PI: Prof Yong-Hua Song (THU) Consortium UK Members: Brunel University
Petascale Software Challenges. William Gropp www.cs.illinois.edu/~wgropp
Petascale Software Challenges William Gropp www.cs.illinois.edu/~wgropp Petascale Software Challenges Why should you care? What are they? Which are different from non-petascale? What has changed since
Mellanox HPC-X Software Toolkit Release Notes
Mellanox HPC-X Software Toolkit Release Notes Rev 1.2 www.mellanox.com NOTE: THIS HARDWARE, SOFTWARE OR TEST SUITE PRODUCT ( PRODUCT(S) ) AND ITS RELATED DOCUMENTATION ARE PROVIDED BY MELLANOX TECHNOLOGIES
Measuring MPI Send and Receive Overhead and Application Availability in High Performance Network Interfaces
Measuring MPI Send and Receive Overhead and Application Availability in High Performance Network Interfaces Douglas Doerfler and Ron Brightwell Center for Computation, Computers, Information and Math Sandia
A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures
11 th International LS-DYNA Users Conference Computing Technology A Study on the Scalability of Hybrid LS-DYNA on Multicore Architectures Yih-Yih Lin Hewlett-Packard Company Abstract In this paper, the
Ultra Low Latency Data Center Switches and iwarp Network Interface Cards
WHITE PAPER Delivering HPC Applications with Juniper Networks and Chelsio Communications Ultra Low Latency Data Center Switches and iwarp Network Interface Cards Copyright 20, Juniper Networks, Inc. Table
ECLIPSE Best Practices Performance, Productivity, Efficiency. March 2009
ECLIPSE Best Practices Performance, Productivity, Efficiency March 29 ECLIPSE Performance, Productivity, Efficiency The following research was performed under the HPC Advisory Council activities HPC Advisory
SR-IOV: Performance Benefits for Virtualized Interconnects!
SR-IOV: Performance Benefits for Virtualized Interconnects! Glenn K. Lockwood! Mahidhar Tatineni! Rick Wagner!! July 15, XSEDE14, Atlanta! Background! High Performance Computing (HPC) reaching beyond traditional
Storage, Cloud, Web 2.0, Big Data Driving Growth
Storage, Cloud, Web 2.0, Big Data Driving Growth Kevin Deierling Vice President of Marketing October 25, 2013 Delivering the Highest ROI Across all Markets HPC Web 2.0 DB/Enterprise Cloud Financial Services
GPI Global Address Space Programming Interface
GPI Global Address Space Programming Interface SEPARS Meeting Stuttgart, December 2nd 2010 Dr. Mirko Rahn Fraunhofer ITWM Competence Center for HPC and Visualization 1 GPI Global address space programming
Advanced Computer Networks. High Performance Networking I
Advanced Computer Networks 263 3501 00 High Performance Networking I Patrick Stuedi Spring Semester 2014 1 Oriana Riva, Department of Computer Science ETH Zürich Outline Last week: Wireless TCP Today:
MPI / ClusterTools Update and Plans
HPC Technical Training Seminar July 7, 2008 October 26, 2007 2 nd HLRS Parallel Tools Workshop Sun HPC ClusterTools 7+: A Binary Distribution of Open MPI MPI / ClusterTools Update and Plans Len Wisniewski
Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand
Exploiting Remote Memory Operations to Design Efficient Reconfiguration for Shared Data-Centers over InfiniBand P. Balaji, K. Vaidyanathan, S. Narravula, K. Savitha, H. W. Jin D. K. Panda Network Based
PMI: A Scalable Parallel Process-Management Interface for Extreme-Scale Systems
PMI: A Scalable Parallel Process-Management Interface for Extreme-Scale Systems Pavan Balaji 1, Darius Buntinas 1, David Goodell 1, William Gropp 2, Jayesh Krishna 1, Ewing Lusk 1, and Rajeev Thakur 1
Software Distributed Shared Memory Scalability and New Applications
Software Distributed Shared Memory Scalability and New Applications Mats Brorsson Department of Information Technology, Lund University P.O. Box 118, S-221 00 LUND, Sweden email: [email protected]
A Hybrid Electrical and Optical Networking Topology of Data Center for Big Data Network
ASEE 2014 Zone I Conference, April 3-5, 2014, University of Bridgeport, Bridgpeort, CT, USA A Hybrid Electrical and Optical Networking Topology of Data Center for Big Data Network Mohammad Naimur Rahman
COMPARISON OF FIXED & VARIABLE RATES (25 YEARS) CHARTERED BANK ADMINISTERED INTEREST RATES - PRIME BUSINESS*
COMPARISON OF FIXED & VARIABLE RATES (25 YEARS) 2 Fixed Rates Variable Rates FIXED RATES OF THE PAST 25 YEARS AVERAGE RESIDENTIAL MORTGAGE LENDING RATE - 5 YEAR* (Per cent) Year Jan Feb Mar Apr May Jun
COMPARISON OF FIXED & VARIABLE RATES (25 YEARS) CHARTERED BANK ADMINISTERED INTEREST RATES - PRIME BUSINESS*
COMPARISON OF FIXED & VARIABLE RATES (25 YEARS) 2 Fixed Rates Variable Rates FIXED RATES OF THE PAST 25 YEARS AVERAGE RESIDENTIAL MORTGAGE LENDING RATE - 5 YEAR* (Per cent) Year Jan Feb Mar Apr May Jun
Performance of RDMA-Capable Storage Protocols on Wide-Area Network
Performance of RDMA-Capable Storage Protocols on Wide-Area Network Weikuan Yu Nageswara S.V. Rao Pete Wyckoff* Jeffrey S. Vetter Ohio Supercomputer Center* Managed by UT-Battelle InfiniBand Clusters around
The Green Index: A Metric for Evaluating System-Wide Energy Efficiency in HPC Systems
202 IEEE 202 26th IEEE International 26th International Parallel Parallel and Distributed and Distributed Processing Processing Symposium Symposium Workshops Workshops & PhD Forum The Green Index: A Metric
Parallel Computing. Benson Muite. [email protected] http://math.ut.ee/ benson. https://courses.cs.ut.ee/2014/paralleel/fall/main/homepage
Parallel Computing Benson Muite [email protected] http://math.ut.ee/ benson https://courses.cs.ut.ee/2014/paralleel/fall/main/homepage 3 November 2014 Hadoop, Review Hadoop Hadoop History Hadoop Framework
Interconnect Your Future Enabling the Best Datacenter Return on Investment. TOP500 Supercomputers, November 2015
Interconnect Your Future Enabling the Best Datacenter Return on Investment TOP500 Supercomputers, November 2015 InfiniBand FDR and EDR Continue Growth and Leadership The Most Used Interconnect On The TOP500
1 st Symposium on Colossal Data and Networking (CDAN-2016) March 18-19, 2016 Medicaps Group of Institutions, Indore, India
1 st Symposium on Colossal Data and Networking (CDAN-2016) March 18-19, 2016 Medicaps Group of Institutions, Indore, India Call for Papers Colossal Data Analysis and Networking has emerged as a de facto
High Performance MPI on IBM 12x InfiniBand Architecture
High Performance MPI on IBM 12x InfiniBand Architecture Abhinav Vishnu Brad Benton Dhabaleswar K. Panda Network Based Computing Lab Department of Computer Science and Engineering The Ohio State University
