The Quantum-ESPRESSO Software Distribution

Similar documents
The Quantum ESPRESSO Software Distribution

Scientific Software Development at Democritos

First Step : Identifying Pseudopotentials

2IP WP8 Materiel Science Activity report March 6, 2013

Ivan Girotto a,, Axel Kohlmeyer a,b,, David Grellscheid a,c, Shawn T. Brown d

The Application of Density Functional Theory in Materials Science

Computer lab: Density functional perturbation theory. theory for lattice dynamics

Part I Courses Syllabus

Q-Chem: Quantum Chemistry Software for Large Systems. Peter M.W. Gill. Q-Chem, Inc. Four Triangle Drive Export, PA 15632, USA. and

Use of openmosix for parallel I/O balancing on storage in Linux cluster

HPC enabling of OpenFOAM R for CFD applications

GPUs for Scientific Computing

ALPS - The Swiss Grand Challenge Programme on the Cray XT3. CUG 2007, Seattle Dominik Ulmer, CSCS

Building a Top500-class Supercomputing Cluster at LNS-BUAP

1 Bull, 2011 Bull Extreme Computing

High Performance Computing. Course Notes HPC Fundamentals

Sourcery Overview & Virtual Machine Installation

BIG CPU, BIG DATA. Solving the World s Toughest Computational Problems with Parallel Computing. Alan Kaminsky

PARALLEL & CLUSTER COMPUTING CS 6260 PROFESSOR: ELISE DE DONCKER BY: LINA HUSSEIN

Graphene a material for the future

GURLS: A Least Squares Library for Supervised Learning

Simulation of Mobile Phone Antenna Performance

COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)

Introduction to Linux and Cluster Basics for the CCR General Computing Cluster

CCTech TM. ICEM-CFD & FLUENT Software Training. Course Brochure. Simulation is The Future

Cloud-WIEN2k. A Scientific Cloud Computing Platform for Condensed Matter Physics

PHonon User s Guide (v. 5.4)

Data Center Specific Thermal and Energy Saving Techniques

Mathematical Libraries on JUQUEEN. JSC Training Course

lesson 1 An Overview of the Computer System

6 J - vector electric current density (A/m2 )

The content is based on the National Science Teachers Association (NSTA) standards and is aligned with state standards.

Science Standard Articulated by Grade Level Strand 5: Physical Science

CE 504 Computational Hydrology Computational Environments and Tools Fritz R. Fiedler

NVIDIA Tesla K20-K20X GPU Accelerators Benchmarks Application Performance Technical Brief

Plate waves in phononic crystals slabs

Parallel Computing with MATLAB

HPC with Multicore and GPUs

for my computation? Stefano Cozzini Which infrastructure Which infrastructure Democrito and SISSA/eLAB - Trieste

Load Balancing on a Non-dedicated Heterogeneous Network of Workstations

Dynamic Load Balancing in CP2K

Performance Improvement of Application on the K computer

Keystone Exams: Chemistry Assessment Anchors and Eligible Content. Pennsylvania Department of Education

Integrated Open-Source Geophysical Processing and Visualization

CS 377: Operating Systems. Outline. A review of what you ve learned, and how it applies to a real operating system. Lecture 25 - Linux Case Study

Next Generation GPU Architecture Code-named Fermi

The Assessment of Benchmarks Executed on Bare-Metal and Using Para-Virtualisation

Use of openmosix for parallel I/O balancing on storage in Linux cluster

Advanced Computational Software

Best practices for efficient HPC performance with large models

Performance analysis of parallel applications on modern multithreaded processor architectures

22S:295 Seminar in Applied Statistics High Performance Computing in Statistics

Nanocomputer & Architecture

Universidad Simón Bolívar

CUDA programming on NVIDIA GPUs

Recommended hardware system configurations for ANSYS users

Hadoop on a Low-Budget General Purpose HPC Cluster in Academia

BIG CPU, BIG DATA. Solving the World s Toughest Computational Problems with Parallel Computing. Alan Kaminsky

Analysis, post-processing and visualization tools

HPC Deployment of OpenFOAM in an Industrial Setting

GPU Hardware and Programming Models. Jeremy Appleyard, September 2015

What is molecular dynamics (MD) simulation and how does it work?

Molecular Dynamics Simulations

Visualization and post-processing tools for Siesta

Defects Introduction. Bonding + Structure + Defects. Properties

The Lattice Project: A Multi-Model Grid Computing System. Center for Bioinformatics and Computational Biology University of Maryland

A Chromium Based Viewer for CUMULVS

ACCELERATING COMMERCIAL LINEAR DYNAMIC AND NONLINEAR IMPLICIT FEA SOFTWARE THROUGH HIGH- PERFORMANCE COMPUTING

CURRICULUM VITAE Andrea Dal Corso

Understanding the Benefits of IBM SPSS Statistics Server

Panasas High Performance Storage Powers the First Petaflop Supercomputer at Los Alamos National Laboratory

MayaVi: A free tool for CFD data visualization

Lecture 2: standard software and visualization tools

64-Bit versus 32-Bit CPUs in Scientific Computing

Linux clustering. Morris Law, IT Coordinator, Science Faculty, Hong Kong Baptist University

Chapter 7: Polarization

Scheduling and Load Balancing in the Parallel ROOT Facility (PROOF)

potential in the centre of the sphere with respect to infinity.

Chapter 3 Student Reading

The Julia Language Seminar Talk. Francisco Vidal Meca

DATA PROCESSING SOFTWARE

QLIKVIEW SERVER LINEAR SCALING

Real Time Analysis of Advanced Photon Source Data

CNR-INFM DEMOCRITOS and SISSA elab Trieste

Robust Algorithms for Current Deposition and Dynamic Load-balancing in a GPU Particle-in-Cell Code

COURSE TITLE COURSE DESCRIPTION

Introduction to bioknoppix: Linux for the life sciences

QUADRICS IN LINUX CLUSTERS

Parallel Simplification of Large Meshes on PC Clusters

RELIABILITY AND NFV BRIAN LEVY CTO SP SECTOR EMEA

Cellular Computing on a Linux Cluster

THE DEVELOPMENT LINES OF PROGEO

Data processing goes big

Designing and Building Applications for Extreme Scale Systems CS598 William Gropp

NMR SPECTROSCOPY. Basic Principles, Concepts, and Applications in Chemistry. Harald Günther University of Siegen, Siegen, Germany.

CHAPTER - 1. Chapter ONE: WAVES CHAPTER - 2. Chapter TWO: RAY OPTICS AND OPTICAL INSTRUMENTS. CHAPTER - 3 Chapter THREE: WAVE OPTICS PERIODS PERIODS

A GENERAL PURPOSE DATA ANALYSIS MONITORING SYSTEM WITH CASE STUDIES FROM THE NATIONAL FUSION GRID AND THE DIII D MDSPLUS BETWEEN PULSE ANALYSIS SYSTEM

Introduction to grid technologies, parallel and cloud computing. Alaa Osama Allam Saida Saad Mohamed Mohamed Ibrahim Gaber

Chap-02, Hardware and Software. Hardware Model

Transcription:

The Quantum-ESPRESSO Software Distribution P. Giannozzi Università di Udine and Democritos National Simulation Center, Italy COST Training School on Molecular and Materials Science GRID applications, Trieste, 2008/09/17 Typeset by FoilTEX

Outline Democritos and scientific software What is Quantum-ESPRESSO? What can Quantum-ESPRESSO do? Quantum-ESPRESSO and the GRID: an experiment with realistic phonon dispersions calculation

Democritos and scientific software The Democritos National Simulation Center, based in Trieste, is dedicated to atomistic simulations of materials, with a strong emphasis on the development of high-quality scientific software Quantum-ESPRESSO is the result of a Democritos initiative, in collaboration with several other institutions (CINECA Bologna, Princeton University, MIT Boston) Quantum-ESPRESSO is a distribution of software for atomistic calculations based on electronic structure, using density-functional theory (DFT), plane-wave (PW) basis set, pseudopotentials (PP) Quantum-ESPRESSO stands for Quantum open-source Package for Research in Electronic Structure, Simulation, and Optimization

Motivations for Quantum-ESPRESSO to merge several pre-existing packages, used by different groups to provide access to several techniques whose usefulness has traditionally been hindered by the limited availability of software: linear response ultrasoft PP/PAW Car-Parrinello Molecular Dynamics to provide a software distribution that is suitable for easy addition of new packages, developments, extensions, features. Such a goal requires: a modular structure, easy to maintain, easily extensible a collaborative environment: free/open-source license

Free software and GNU license GNU (Gnu s Not Unix) General Public License (GPL) is probably the most common free-software license. Basically: The source code is available. You can do whatever you want with the sources, but if you distribute any derived work, you have to distribute under the GPL the sources of the derived work. Advantages: Everybody including commercial entities can contribute. Nobody can steal the code and give nothing back to the community.

Quantum-ESPRESSO as a distribution Quantum-ESPRESSO aims at becoming a distribution of packages, rather than a single, monolithic, tightly integrated package. Main packages presently in the Quantum-ESPRESSO distribution: PWscf: self-consistent electronic structure, structural optimization, dynamics, linear-response calculations (phonons, dielectric properties), ballistic conductance CP/FPMD: variable-cell Car-Parrinello molecular dynamics They share a common installation method, input format, PP format, data output format, large parts of the basic code.

Quantum-ESPRESSO as a distribution (2) Additional packages: GIPAW: Gauge-Independent PAW method for EPR and NMR chemical shifts W90: Maximally Localised Wannier functions atomic: auxiliary code for PP generation PWGui: a Graphical User Interface for production of input files plus other auxiliary codes for data postprocessing. Coming soon: XSpectra: simulation of X-ray adsorption spectra More packages planned for addition.

Organization The distribution is maintained as a single CVS (Concurrent Version System) tree. Available to everyone anytime via anonymous access. Web sites: http://www.quantum-espresso.org (for Quantum-ESPRESSO in general) http://www.pwscf.org (specific to PWscf) Mailing list: pw users: used by developers for announcements about Quantum- ESPRESSO pw forum: for general discussions (all subscribed users can post)

Technical characteristics (coding) mostly written in Fortran-90, with various degrees of sophistication (i.e. use of advanced f90 features) no dusty decks any longer use of standard library routines (lapack, blas, fftw) to achieve portability Machine-optimized libraries can be used if available C-style preprocessing options allow to keep a single source tree for all architectures various levels of parallelization via MPI calls, hidden into calls to very few routines (almost) unified serial and parallel versions; parallel code can (usually) be written without knowing the details Easy (or not-so-difficult) installation via the GNU utility configure

What can Quantum-ESPRESSO do? Quantum-ESPRESSO can perform structural optimization and find the corresponding ground-state electronic structure in molecules, clusters, surfaces, perfect and defective crystals insulators and metals magnetic as well as nonmagnetic systems Quantum-ESPRESSO can perform more complicated calculations such as e.g. finding transition states and minimum-energy pathways for chemical reactions (next two slides courtesy of Stefano de Gironcoli, SISSA)

Quantum-ESPRESSO can perform first-principle molecular dynamics. An example: reconstruction of the surface of Ice with an excess electron (courtesy of Sandro Scandolo, ICTP)

Quantum-ESPRESSO can deal with complicated organometallic compounds and models of active sites in proteins. An example: 332-atom model of a heme site, including all atoms in a sphere within 8 Å radius centered around the Iron atom of the porphyrin PG, F. de Angelis, R. Car, JCP 120, 8632 (2004)

Computer requirements Quantum-ESPRESSO is both CPU and RAM-intensive. Actual CPU time and RAM requirements depend upon: size of the system under examination: CPU N 2 3, RAM N 2, where N = number of atoms in the supercell or molecule kind of system: type and arrangement of atoms, influencing the number of PWs, of electronic states, of k-points needed desired results: increasing effort from simple self-consistent (singlepont) calculation to structural optimization to reaction pathways, molecular-dynamics simulations CPU time mostly spent in FFT and linear algebra. RAM mostly needed to store wavefunctions (Kohn-Sham orbitals)

Typical computational requirements Basic step: self-consistent ground-state DFT electronic structure. Simple crystals, small molecules, up to 50 atoms CPU seconds to hours, RAM up to 1-2 Gb: may run on single PC Surfaces, larger molecules, complex or defective crystals, up to a few hundreds atoms CPU hours to days, RAM up to 10-20 Gb: requires PC clusters or conventional parallel machines Complex nanostructures or biological systems CPU days to weeks or more, RAM tens to hundreds Gb: massively parallel machines Excessive RAM requirements for single PCs leave no choice, but main factor pushing towards parallel machines is excessive CPU time.

Quantum-ESPRESSO and High-Performance Computing A considerable effort has been devoted to Quantum-ESPRESSO parallelization. Several parallelization levels are implemented; the most important, PW parallelization, requires fast communications. Recent achievements (mostly due to Carlo Cavazzoni, CINECA): porting to the BlueGene architecture (BG/L and BG/P) CP has been run on up to 4800 Cray XT4 processors, reaching 10 TFlops, for a 1532-atom porphyrin-functionalized nanotube obtained via addition of more parallelization levels and via careful optimization of nonscalable RAM and computations.

Quantum-ESPRESSO and the GRID Large-scale computations with Quantum-ESPRESSO require large parallel machines with fast communications: unsuitable for GRID. BUT: often many smaller-size, loosely-coupled or independent computations are required. A few examples: calculations under different conditions (pressure, temperature) or for different compositions, or for different values of some parameters; the search for materials having a desired property (e.g. bulk modulus); largest transition path search (NEB); full phonon dispersions in crystals

Hand-made GRID computing

Phonons in crystals Phonon frequencies ω(q) are determined by the secular equation: C αβ st (q) M s ω 2 (q)δ st δ αβ =0 where C αβ st (q) is the matrix of force constants for a given q

Calculation of phonon dispersions αβ The force constants C st (q) are calculated for a uniform grid of n q q-vectors, then Fourier-transformed to real space For each of the n q q-vectors, one has to perform 3N linearresponse calculations, one per atomic polarization; or equivalently, 3ν calculations, one per irrep (symmetrized combinations of atomic polarizations, whose dimensions range from 1 to 4) Grand total: 3νn q calculations, may easily become heavy. But: Each C αβ st (q) matrix is independently calculated, then collected Each irrep calculation is almost independent except at the end, when the contributions to the force constant matrix are collected Perfect for execution on the GRID!

Technicalities Only minor changes needed in the phonon code, namely possibility to run one q-vector at the time (already there) possibility to run one irrep at the time and to save partial results (a single row of the force constant matrix) Server-client application (written by Riccardo di Meo) takes care of dispatching jobs and of collecting results

A realistic phonon calculation on the GRID γ-al 2 O 3 has a (simplified) unit cell of 40 atoms, i.e. 120 n q linear-response calculations, n q 10: a few weeks on a single PC. Currently running on a lot of machines!

Credits Thanks to the many people have contributed to Quantum- ESPRESSO to Stefano Cozzini for arising in me the interest in GRID computing with Quantum-ESPRESSO to Riccardo di Meo, Riccardo Mazzarello and Andrea Dal Corso who did the real work to Eduardo Ariel Menendez Proupin (Santiago, Chile) who suggested phonons in γ-al 2 O 3...and thank you for your attention!