The Extension of SuperMUC: Phase 2 - Inside

Editorial
Editorial
Publishers
Editor
Design
2
3
Contents
Contents
Contents
News
Applications
Systems
Centres
Activities
Projects
Courses
News
News
PRACE: Results of the
7th Regular Call
PRACE Projects meet in
Varna
News
News
Interview with
Prof. Dr.-Ing. Dr. h.c. Dr. h.c.
Michael M. Resch
News
News
Prof. Michael M. Resch is the
Chairman of the Board of Directors
of GCS since May 2013. He was a
cofounder of GCS. Prof. Resch is
currently the director of the High
Performance Computing Center
Stuttgart (HLRS), the director of the
Information Service Center (IZUS)
of the University of Stuttgart and
the director of the Institute for
High Performance Computing (IHR)
of the University of Stuttgart.
Applications
Applications
Simulating the Life Cycle
of molecular Clouds
2
2
Numerical Method
Applications
Applications
The typical Milky Way Disk
2
2
2
2
2
Applications
Applications
Disk Galaxies at different
Gas Surface Densities: from
low to high Redshift
References
[1]
2
2
Conclusions and Outlook
Bigiel, F., Leroy, A., Walter, F., Brinks,
E., de Blok, W. J. G., Madore, B.,
Thornley, M. D.
[2] Brown, P. N., Byrne, G. D.,
Hindmarsh, A. C.
[3] Chabrier, G.
[4] Clark, P. C., Glover, S. C. O.,
Klessen, R. S.
2
[5] Evans, II, N. J. and et al.
[6] Fryxell, B., et al.
[7]
2
Glover, S. C. O., Mac Low, M.-M.
[8] Górski, K. M., Hivon, E.
2
2
3
2
[9] Mac Low, M.-M., Klessen, R. S.
2
2
[10] Marinacci, F., Fraternali, F., Nipoti, C.,
Binney, J., Ciotti, L., Londrillo, P.
2
[11] Moster, B. P., Somerville, R. S.,
Maulbetsch, C., van den Bosch, F. C.,
Macciò, A. V., Naab, T., Oser, L.
[12] Newman, S. F., et al.
2
[13] Tacconi, L., et al.
3
Applications
Applications
Gadget3: Numerical
Simulation of Structure
Formation in the Universe
3
Applications
Applications
References
[1]
Springel, V.
[2] Springel, V., Yoshida, N., White, S.D.M.
2
2
Applications
Applications
Numerical Simulation of
Correlated Electron Systems
The Kane-Mele-Hubbard
Model
The Numerical Challenge
22
23
Applications
Applications
Results
Impact of U for
2
=0
Applications
Applications
[11] Hohenadler, M., Meng, Z. Y., Lang, T. C.,
Wessel, S., Muramatsu, A., Assaad, F. F.
[12] Berg, E., Metlitski, M.A., Sachdev, S.
Acknowledgments
[13] Rachel, S., Le Hur, K.
[14] Assaad, F.F., Bercx, M., Hohenadler, M.
[15] Herbut, I.F.
Impact of U for
>0
[16] Herbut, I.F., Juri i
V., Vafek, O.
[17] Yang, H.-Y., Albuquerque, A.F.,
Capponi, S., Läuchli, A.M., Schmidt, K.P.
References
[1]
[18] Chang, C.-C., Scalettar, R.T.
Kane, C.L., Mele, E.J.
[19] Clark, B.K.
[2] König, M., Wiedmann, S., Brüne, C.,
Roth, A., Buhmann, H., Molenkamp, L.W.,
Qi, X.-L., Zhang, S.-C.
[3] Meng, Z.Y., Lang, T.C., Wessel, S.,
Assaad, F.F., Muramatsu, A.
[20] Sorella, S., Tosatti, E.
[21] Paiva, T., Scalettar, R.T., Zheng, W.,
Singh, R.R.P., Oitmaa, J.
[22] Qi, X.-L., Zhang, S.-C.
[4] Sorella, S., Otsuka, Y., Yunoki, S.
[23] Ran, Y., Vishwanath, A., Lee, D.-H.
[5] Assaad, F.F., Herbut, I.F.
[6] Hohenadler, M., Assaad, F.F.
[7]
Outlook
Uehlinger, T., Jotzu, G., Messer, M.,
Greif, D., Hofstetter, W., Bissbort, U.,
Esslinger, T.
[8] Assaad, F.F., Evertz, H.G.
[9] Assaad, F.F.
[10] Hohenadler, M., Lang, T.C., Assaad, F.F.
Applications
Applications
Highly-resolved numerical
Simulations of bed-load Transport
in a turbulent open Channel Flow
Numerical Method
Computational Setup
Results
Applications
Applications
References
[1]
Dietrich, W.E., Kirchner, J.W., Ikeda,
H., Iseya, F.
[2] Shvidchenko, A.B., Pender, G.
[3] Kempe, T., Fröhlich, J.,
[4] Kempe, T., Fröhlich, J.,
[5] Vowinckel, B., Kempe, T., Fröhlich, J.,
Conclusions
2.5
1
0.8
Fix
Ref
FewPart
LowSh
Ref
FewPart
LowSh
2
1.5
Gm( x,0)
Ruu
0.6
0.4
0.2
[7]
Yalin, M.S., Ferreira da Silva, A.M.,
[8] Vowinckel, B., Kempe, T., Fröhlich, J.,
Nikora, V.I.
1
0.5
0
-0.2
[6] Shields, A.
0
2
4
6
rx / H
8
10
12
0
0
2
4
x
6
/H
8
10
12
[9] Vowinckel, B., Kempe, T., Fröhlich, J.
Applications
Applications
How to fit the Local Universe
into a Supercomputer?
Recovering and Simulating
Structures of the Local
Universe
32
33
Applications
Applications
Acknowledgements
References
[1]
Heß, S., Kitaura, F.-S., Gottlöber, S.
[2] Kitaura, F.-S.
[3] Kitaura, F.-S., Heß, S.
Applications
Applications
A scalable hybrid DFT/PMM-MD
Approach for accurately simulating Biomolecules on SuperMUC
Applications
Applications
References
[1]
Senn, H.M., Thiel, W.
[2] Schwörer, M., Breitenfeld, B.,
Tröster, P., Bauer, S., Lorenzen, K.,
Tavan, P., Mathias, G.
[3] Lorenzen, K., Schwörer, M., Tröster, P.,
Mates, S., Tavan, P.
[4]
[5] Mathias, G., Baer, M.D.
2
Applications
Applications
Aircraft Wake Vortex
Evolution during Approach
and Landing
With and without Plate Lines
Applications
Applications
3
0
3
Applications
Applications
References
[1]
Holzäpfel, F., Steen, M.
[2] Holzäpfel, F., Gerz, T., Frech, M.,
Tafferner, A., Köpp, F., Smalikho, I.,
Rahm, S., Hahn, K.-U., Schwarz, C.
[3] Misaka, T., Holzäpfel, F., Gerz, T.
[4] Stephan, A., Holzäpfel, F., Misaka, T.
Links
Projects
Projects
FORTISSIMO
Factories of the Future
Resources, Technology, Infrastructure and Services for
Simulation and Modelling
Projects
Projects
References
[1] Sawyer, P.
[2] Schubert, J.
[3]
Core Project Partners
Projects
Projects
Revisiting Dynamic Scheduling
Techniques for HPC Infrastructures: The Approach
of the DreamCloud Project
[3]
Project Partners
References
[1]
[2]
[3]
[4]
[1]
[2]
[4]
Projects
Projects
SkaSim - Scalable HPC-Codes
for molecular Simulation in the
Chemical Industry
Projects
Projects
POLCA: Programming
Large Scale Heterogeneous
Infrastructures
Projects
Projects
The POLCA Approach
Programming with POLCA
Projects
Projects
What POLCA will provide
Who is POLCA?
Projects
Projects
A flexible Framework for Energy
and Performance Analysis
of highly parallel Applications
in a Supercomputing Centre
Background Software
Projects
Projects
Summary
References
[1]
[2]
[3] Focht, E., Jeutter, A.
[4] Guillen, C., Hesse, W., Brehm, M.
[5]
[6] Treibig, J., Hager, G., Wellein, G.
2
Focus of the
Project Implementation
[7]
[8]
[9]
Projects
Projects
SIMOPEK – Simulation and
Optimization of Data Center
Energy Flows from Cooling
Networks taking into
Account HPC Operation
Scenarios
External Influences/Constraints
Data Center: Reduce Total Cost of Ownership
Pillar 1
Building Infrastructure
Pillar 2
HPC System Hardware
Pillar 3
HPC System Software
System Scheduler
Infrastructure Aware Resource Management & Scheduling
U!lity Providers
Neighboring Buildings
Advanced Heat Reuse Technologies
Pillar 4
HPC Applications
Modeling, Simula!on & Op!miza!on
Data Center Data Acquisi!on Monitor
Building
Management &
Infrastructure
Hardware
Management
System
Management
So"ware
Performance
Analysis Tools
Infrastructure
Monitoring
System Hardware
Monitoring
System So"ware
Monitoring
Performance
Monitoring
Projects
Projects
References
[1]
Wilde, T., Auweter, A., Shoukourian, H.
[2] Shoukourian, H., Wilde, T., Auweter, A.
External Influences/Constraints
[3]
Data Center: Reduce Total Cost of Ownership
Pillar 1
Building Infrastructure
Pillar 3
HPC System Software
Pillar 2
HPC System Hardware
SIMOPEK Power Consump!on Modeling,
Simula!on & Op!miza!on using MYNTS
Facts and Figures
FEPA Performance, Energy Modeling & Op!miza!on
FEPA Data Collec!on
SIMOPEK Data Collec!on using PowerDam V.2.0
Building
Management &
Infrastructure
Hardware
Management
[4]
System Scheduler
Infrastructure Aware Resource Management & Scheduling
U!lity Providers
Neighboring Buildings
SIMOPEK Advanced Absorp!on Cooling
Pillar 4
HPC Applications
System
Management
So"ware
2
Performance
Analysis Tools
2
Infrastructure
Monitoring
System Hardware
Monitoring
System So"ware
Monitoring
Performance
Monitoring
Projects
Projects
The Catwalk Project –
A quick Development Path
for Performance Models
Projects
Projects
References
[1]
Calotoiu, A., Hoefler, T., Poke, M.,
Wolf, F.
[2] Geimer, M., Wolf, F., Wylie, B.J.N.,
Ábrahám, E., Becker, D., Mohr, B.
[3] an Mey, D., Biersdorff, S., Bischof, C.,
Diethelm, K., Eschweiler, D., Gerndt, M.,
Knüpfer, A., Lorenz, D., Malony, A.D.,
Nagel, W.E., Oleynik, Y., Rössel, C.,
Saviankou, P., Schmidl, D., Shende, S.S.,
Wagner, M., Wesarg, B., Wolf, F.
Projects
Projects
GROMEX – Unified Longrange Electrostatics and
Flexible Ionization
Usability & Scalability
Background of the Project
Towards realistic Simulations
Projects
Projects
Project Partners
References
(1)
Kabadshow, I., Dachsel, H.
(4) Hess, B., Kutzner, C., van der Spoel, D.,
Lindahl, E.
(5) Ullmann, R.T., Ullmann, G.M.
(2) Dachsel, H.
(3) Donnini, S., Tegeler, F., Groenhof, G.,
Grubmüller, H.
2
Projects
Projects
HOPSA – A big Jump
forward in HPC System
and Application Monitoring
Integration among the
HOPSA Performance
Analysis Tools
The HOPSA Performance
Tool Workflow
Projects
Projects
Conclusion
Integration of System Data
and Performance Analysis
Tools
Projects
Projects
[5] an Mey, D., Biersdorff, S., Bischof, C.,
Diethelm, K., Eschweiler, D., Gerndt, M.,
Knüpfer, A., Lorenz, D., Malony, A.D.,
Nagel, W.E., Oleynik, Y., Rössel, C.,
Saviankou, P., Schmidl, D., Shende, S.S.,
Wagner, M., Wesarg, B., Wolf, F.
Russian Project Partners
(HOPSA-RU)
[6] Servat, H., Llort, G., Giménez, J.,
Labarta, J.
EU Project Partners
(HOPSA-EU)
[7]
References
[1]
Labarta, J., Girona, S., Pillet, V.,
Cortes, T., Gregoris, L.
[2] Geimer, M., Wolf, F., Wylie, B.J.N.,
Abraham, E., Becker, D., Mohr, B.
[8] Adinets, A.V., Bryzgalov, P.A., Vad, V.,
Voevodin, V., Zhumatiy, S.A.,
Nikitenko, D.A.
[9] Mohr, B., Voevodin, V., Giménez, J.,
Hagersten, E., Knüpfer, A.,
Nikitenko, D.A., Nilsson, M., Servat, H.,
Shah, A., Winkler, F., Wolf, F., Zhujov, I.
[3] Berg, E., Hagersten, E.
[10]
[4] Nagel, W., Weber, M., Hoppe, H.-C.,
Solchenbach, K.
Systems
Systems
End of the HPC-FF Era
Systems
Systems
2
References
[1]
[2]
[3]
[4]
Systems
Systems
JUROPA-3 - A Prototype for
the Next-Generation HPC
Cluster
System Specifications
Systems
Systems
First Experiences with
the Intel MIC Architecture
at LRZ
Architectural Overview
Intel MIC Architecture
at LRZ
Programming Models
Number of cores
Frequency of cores
GDDR5 memory size
Number of hardware threads
SIMD vector registere
Flops/cycle
Theoretical peak performance
L2 cache per core
Systems
Systems
Benchmarks
Acknowledgements
References
(1)
(2)
(3)
Weinberg, V., (Editor) et al.
Systems
Systems
The Extension of
SuperMUC: Phase 2
SuperMUC – Phase 1
Users from 25 European
Countries
Phase 2
Innovative Water Cooling
Financing
References
(1)
(2)
(3)
Centres
Centres
Leibniz Supercomputing Centre of
the Bavarian Academy of Sciences and
Humanities (Leibniz-Rechenzentrum, LRZ)
provides comprehensive services to
scientific and academic communities by:
Research in HPC is carried out in collaboration with the distributed, statewide
Competence Network for Technical and
Scientific High Performance Computing
in Bavaria (KONWIHR).
Contact:
Leibniz Supercomputing Centre
Prof. Dr. Arndt Bode
Boltzmannstr. 1
85478 Garching near Munich
Germany
Phone +49 - 89 - 358 - 31- 80 00
[email protected]
www.lrz.de
Compute servers currently operated by LRZ are given in
the following table
System
Size
Peak
Performance
(TFlop/s)
Purpose
User Community
Centres
Centres
First German National Center
Based on a long tradition in supercomputing at University of Stuttgart, HLRS
(Höchstleistungsrechenzentrum Stuttgart)
was founded in 1995 as the first German
federal Centre for High Performance
Computing. HLRS serves researchers
at universities and research laboratories
in Europe and Germany and their external and industrial partners with high-end
computing power for engineering and
scientific applications.
Service for Industry
Service provisioning for industry is done
together with T-Systems, T-Systems sfr,
and Porsche in the public-private joint
venture hww (Höchstleistungsrechner
für Wissenschaft und Wirtschaft).
Through this co-operation industry
always has acces to the most recent
HPC technology.
Bundling Competencies
In order to bundle service resources in
the state of Baden-Württemberg HLRS
has teamed up with the Steinbuch Center for Computing of the Karlsruhe Institute of Technology. This collaboration
has been implemented in the non-profit
organization SICOS BW GmbH.
World Class Research
As one of the largest research centers
for HPC HLRS takes a leading role in
research. Participation in the German
national initiative of excellence makes
HLRS an outstanding place in the field.
Contact:
Höchstleistungsrechenzentrum
Stuttgart (HLRS)
Universität Stuttgart
Prof. Dr.-Ing. Dr. h.c. Dr. h.c.
Michael M. Resch
Nobelstraße 19
70569 Stuttgart
Germany
Phone +49 - 711- 685 - 8 72 69
[email protected] / www.hlrs.de
Hermit
Compute servers currently operated by HLRS
Peak
Performance
User
Community
System
Size
(TFlop/s)
Purpose
Cray XE6
"Hermit"
(Q4 2011)
3,552 dual socket
nodes with 113,664
AMD Interlagos cores
1,045
Capability
Computing
European
and German
Research
Organizations
and Industry
NEC Cluster
(Laki, Laki2)
heterogenous
compunting
platform of 2
independent
clusters
23 TB memory
9988 cores
911 nodes
170
Laki:
120,5 TFlops
Laki2:
47,2 TFlops
German
Universities,
Research
Institutes
and Industry
Centres
Centres
Supercomputer-oriented research
and development in selected fields
of physics and other natural sciences
by research groups of competence in
supercomputing applications.
The Jülich Supercomputing Centre (JSC)
at Forschungszentrum Jülich enables
scientists and engineers to solve grand
challenge problems of high complexity in
science and engineering in collaborative
infrastructures by means of supercomputing and Grid technologies.
Provision of supercomputer resources
of the highest performance class for projects in science, research and industry
in the fields of modeling and computer
simulation including their methods. The
selection of the projects is performed
by an international peer-review procedure
implemented by the John von Neumann
Institute for Computing (NIC), a joint
foundation of Forschungszentrum Jülich,
Deutsches Elektronen-Synchrotron
DESY, and GSI Helmholtzzentrum für
Schwerionenforschung.
Implementation of strategic support
infrastructures including communityoriented simulation laboratories and
cross-sectional groups on mathematical
methods and algorithms and parallel
performance tools, enabling the effective usage of the supercomputer
resources.
Higher education for master and doctoral students in cooperation e.g.
with the German Research School for
Simulation Sciences.
Contact:
Jülich Supercomputing Centre (JSC)
Forschungszentrum Jülich
Prof. Dr. Dr. Thomas Lippert
52425 Jülich
Germany
Phone +49 - 24 61- 61- 64 02
[email protected]
www.fz-juelich.de/jsc
Compute servers currently operated by JSC
Peak
Performance
System
Size
(TFlop/s)
®
Purpose
User
Community
Activities
Activities
CECAM Tutorials at JSC
2
CHANGES Workshop
Activities
Activities
Laboratory Experiments
on Crowd Dynamics
Activities
Activities
JSC Guest Student
Programme on Scientific
Computing 2013
Activities
Activities
High-Q Club – The highest
scaling Codes on JUQUEEN
Terra-Neo
Gysela
waLBerla
PEPC
PMG+PFASST
Reference
[1]
dynQCD
Brömmel, D.
Activities
Activities
Jülich Supercomputing
Centre contributes to visionary
“Human Brain Project”
References
[1]
[2]
[3]
Activities
Activities
Traffic and Granular Flow
Conference celebrates 10th
Edition by returning to Jülich
UNICORE Summit 2013
References
[1]
[2]
Reference
[1]
Activities
Activities
3D Show at the Pharma Forum:
Simulation and Visualization of
the Airflow in Cleanrooms
Links
s
Activities
Activities
The 17th HLRS-NEC
Workshop on Sustained
Simulation Performance
s
Activities
Activities
ls1 mardyn - a Massively Parallel
Molecular Simulation Code
Scalability
Neighbour Search
Dynamic Load Balancing
Partners
s
s
s
s
s
s
Activities
Activities
GCS at ISC’13 – Review
GCS Booth Highlights
2
s
Two GCS HPC systems
amongst Top Ten of TOP500
ISC’13 Gauss Award Winner
Activities
Activities
Extreme Scaling
Workshop at LRZ
July 9-11, 2013:
Running Real World
Applications on more than
130,000 Cores on SuperMUC
s"1#$ISLANDS
s#)!/ISLANDS
s6ERTEXISLANDS
s0'ADGET88,ISLANDS
s'2/-!#3ISLANDS
s,!--03ISLANDS
s.YXISLANDS
Activities
Activities
sWAL"%RLAISLANDS
s!0%3ISLANDS
LRZ Extreme Scale Benchmark and Optimization Suite
s%XA-,ISLANDS
s)#/.ISLANDS
3
IBM MPI 1.3 + icc 12.1
Intel MPI 4.1 + icc 13.1
81,000 atoms aquaporin, PME, 2 fs
2 M atoms ribosome, PME, 4 fs
12 M atoms peptides, PME, 2 fs
s3EIS3OLISLANDS
Performance Results
Activities
Activities
HLRS Scientific Tutorials and
Workshop Report and Outlook
OpenACC Programming for Parallel Accelerated
Supercomputers – an alternative to
CUDA from Cray perspective
Cray XE6/
XC30 Optimization Workshops
PRACE Advanced Training Centre
Parallel Programming Workshop
2014 – Workshop Announcements
Iterative Solvers and
Parallelization
Scientific Conferences and Workshops at HLRS
12th HLRS/hww Workshop on Scalable Global Parallel File Systems (March/April 2014)
8th ZIH+HLRS Parallel Tools Workshop (date and location not yet fixed)
High Performance Computing in Science and Engineering - The 17th Results and
Review Workshop of the HPC Center Stuttgart (October 2014)
IDC International HPC User Forum (October 2014)
Parallel Programming Workshops: Training in Parallel Programming and CFD
Parallel Programming and Parallel Tools (TU Dresden, ZIH, February 24 - 27)
Introduction to Computational Fluid Dynamics (HLRS, March 31 - April 4)
ISC and SC Tutorials
Iterative Linear Solvers and Parallelization (HLRS, March 24-28)
Georg Hager, Gabriele Jost, Rolf Rabenseifner: Hybrid Parallel Programming with MPI & OpenMP. Tutorial 9 at the International Supercomputing
Conference, ISC’13, Leipzig, June 16-20. 2013.
Georg Hager, Jan Treibig, Gerhard Wellein: Node-Level Performance
Engineering. Tutorial 2 at the International Supercomputing Conference,
ISC’13, Leipzig, June 16-20. 2013.
Rolf Rabenseifner, Georg Hager, Gabriele Jost: Hybrid MPI and OpenMP
Parallel Programming. Half-day Tutorial at Super Computing 2013,
SC13, Denver, Colorado, USA, November 17-22, 2013.
Cray XE6/XC30 Optimization Workshops (HLRS, March 17 - 20) (PATC)
GPU Programming using CUDA (HLRS, April 7 - 9)
Introduction
to Computational Fluid Dynamics
Open ACC Programming for Parallel Accelerated Supercomputers
(HLRS, April 10 - 11) (PATC)
MPI & OpenMP
Unified Parallel C (UPC) and Co-Array Fortran (CAF) (HLRS, April 14 - 15) (PATC)
Scientific Visualisation (HLRS, April 16 - 17)
Fortran
for Scientific Computing
Parallel Programming with MPI & OpenMP (TU Hamburg-Harburg, July 28 - 30)
Iterative Linear Solvers and Parallelization (LRZ, Garching, September 15 - 19)
Introduction to Computational Fluid Dynamics (ZIMT Siegen, September/October)
Message Passing Interface (MPI) for Beginners (HLRS, October 6 - 7) (PATC)
Shared Memory Parallelization with OpenMP (HLRS, October 8) (PATC)
Advanced Topics in Parallel Programming (HLRS, October 9 - 10) (PATC)
Parallel Programming with MPI & OpenMP (FZ Jülich, JSC, December 1 - 3)
Training in Programming Languages at HLRS
Fortran for Scientific Computing (Dec 2 - 6, 2013 and Mar 10 - 14, 2014) (PATC)
URLs:
Unified Parallel C (UPC) and CoArray Fortran (CAF)
http://www.hlrs.de/events/
http://www.hlrs.de/training/course-list/
(PATC): This is a PRACE PATC course
GCS – High Performance Computing
Parallel Programming with
MPI, OpenMP and PETSc
Date and Location
model, compilers, tools, monitoring,
computational science, and also
sessions will allow users to immediately
Contents
MPI, OpenMP, performance optimi-
show how an awareness of the per-
test and understand the language
This course is targeted at scientists
zation, mathematical software, and
formance features of an application
constructs.
with little or no knowledge of the
application software.
may lead to notable reductions in
Contents
Fortran programming language, but
Date and Location
Web Page
needing it for participation in projects
February 24 - 27, 2014
Web Page
s )NTRODUCTION
www.hlrs.de/training/course-list
using a Fortran code base, for devel-
Dresden, ZIH
http://www.fz-juelich.de/ias/jsc/
s 0RACTICALPERFORMANCEANALYSIS
events/sc-nov
s -ICROBENCHMARKSANDTHEMEMORY
The focus is on programming models
hierarchy
s 4YPICALNODELEVELSOFTWAREOVER
sessions (in C and Fortran) will allow
Node-Level Performance
Engineering
users to immediately test and
(PATC course)
s %XAMPLEPROBLEMS
MPI, OpenMP, and PETSc. Hands-on
understand the basic constructs of
Contents
tools like debugger and syntax checker
The focus is on programming models
(PATC course)
as well as handling of compilers and
MPI, OpenMP, and PETSc. Hands-on
libraries. The language is for the most
sessions (in C and Fortran) will allow
part treated at the level of the Fortran
users to immediately test and
- The 3D Jacobi solver
February 03 - 05, 2014
95 standard; features from Fortran
understand the basic constructs of
JSC, Forschungszentrum Jülich
2003 are limited to improvements on
the Message Passing Interface (MPI)
- The Lattice-Boltzmann Method
and the shared memory directives of
December 03 - 04, 2013
- Sparse Matrix-Vector Multiplication
OpenMP. Course language is English.
LRZ Building,
- Backprojection algorithm for CT
This course is organized by JSC in
University Campus Garching,
collaboration with HLRS. Presented
near Munich, Boltzmannstr. 1
Contents
getting acquainted with additional
Date and Location
Date and Location
by Dr. Rolf Rabenseifner, HLRS.
opment of their own codes, and for
Second JUQUEEN Porting
and Tuning Workshop
heads
the Message Passing Interface (MPI)
the elementary level. Advanced
and the shared memory directives of
Contents
Fortran features like object-oriented
OpenMP. The last day is dedicated to
The Blue Gene/Q petaflop super-
programming or coarrays will be cov-
tools for debugging and performance
s %NERGY0ARALLEL3CALABILITY
computer JUQUEEN marks another
ered in a follow-on course in autumn.
analysis of parallel applications. This
Between each module, there is time
quantum leap in supercomputer
To consolidate the lecture material,
course is organized by ZIH in collabo-
for Questions and Answers!
performance at JSC. In order to use
each day's approximately 4 hours of
ration with HLRS.
this tool efficiently, special efforts by
lecture are complemented by 3 hours
of hands-on sessions.
reconstruction
Web Page
This course teaches performance
http://www.fz-juelich.de/ias/jsc/
engineering approaches on the
Web Page
the users are necessary, though. The
events/mpi
compute node level. “Performance
http://www.lrz.de/services/
aim of this hands-on workshop is to
engineering” as we define it is more
compute/courses
support current users of JUQUEEN
Prerequisites
in porting their software, in analyzing
Course participants should have ba-
its performance, and in improving
sic UNIX/Linux knowledge (login with
developing a thorough understanding
Fortran for Scientific
Computing
its efficiency. This course is a PATC
secure shell, shell commands, basic
of the interactions between software
(PATC course)
course (PRACE Advanced Training
programming, vi or emacs editors).
Introduction to the
Programming and Usage
of the Supercomputer
Resources at Jülich
Parallel Programming with
MPI, OpenMP, and Tools
power consumption:
November 25 - 27, 2013
JSC, Forschungszentrum Jülich
Courses and Tutorials
than employing tools to identify
hotspots and bottlenecks. It is about
and hardware. This process must
Web Page
www.hlrs.de/training/course-list
Parallel Programming of
High Performance Systems
Dates and Location
March 10 - 14, 2014
Centres).
Date and Location
start at the core, socket, and node
Dates and Location
Web Page
RRZE building, University campus
November 28 - 29, 2013
level, where the code gets executed
December 02 - 06, 2013 and
JSC, Forschungszentrum Jülich
that does the actual computational
March 10 - 14, 2014
Web Page
http://www.lrz.de/services/com-
Erlangen, Martensstr. 1: Via video
http://www.fz-juelich.de/ias/jsc/
pute/courses
conference at LRZ if there is sufficient
work. Once the architectural require-
Stuttgart, HLRS
events/juqueenpt14
interest.
Programming with Fortran
Contents
Contents
ments of a code are understood and
This course gives an overview of
correlated with performance mea-
Contents
the supercomputers JUROPA and
surements, the potential benefit of
This course is dedicated for scientists
JUQUEEN. Especially new users will
optimizations can often be predicted.
and students to learn (sequential)
Dates and Locations
Erlangen Regional Computing Centre
learn how to program and use these
We introduce a “holistic” node-level
programming scientific applications
February 03 - 07, 2014
(RRZE) and LRZ, is targeted at stu-
systems efficiently. Topics discussed
performance engineering strategy,
with Fortran. The course teaches
LRZ Building, University campus
dents and scientists with interest in
are: system architecture, usage
apply it to different algorithms from
newest Fortran standards. Hands-on
Garching near Munich.
programming modern HPC hardware,
This course, a collaboration of
GCS – High Performance Computing
specifically the large scale parallel
on these opportunities. From Monday
and the shared memory directives of
computing systems available in Munich,
to Wednesday, specialists from Cray
OpenMP. This course is organized by
Jülich and Stuttgart.
will support you in your effort porting
University of Kassel, HLRS, and IAG.
and optimizing your application on our
Courses and Tutorials
Introduction to Computational Fluids Dynamics
Contents
In this add-on course to the parallel
GPU Programming using
CUDA
programming course special topics
Date and Location
are treated in more depth, in
Date and Location
Each day is comprised of approxi-
Cray XE6. On Thursday, Georg Hager
Web Page
March 31 - April 04, 2014
particular performance analysis, I/O
April 07 - 09, 2014
mately 4 hours of lectures and 3
and Jan Treibig from RRZE will present
www.hlrs.de/training/course-list
Stuttgart, HLRS
and PGAS concepts. It is provided
Stuttgart, HLRS
hours of hands-on sessions.
detailed information on optimizing
Eclipse: C/C++/Fortran
programming
Contents
Computing Centre (RRZE) and LRZ
Contents
Numerical methods to solve the
within KONWIHR.
The course provides an introduction
equations of Fluid Dynamics are pre-
Each day is comprised of approxima-
to the programming language CUDA,
codes on the multicore AMD Interlagos
Web Page
processor. Course language is English
http://www.lrz.de/services/
(if required).
Date and Location
sented. The main focus is on explicit
tely 5 hours of lectures and 2 hours
which is used to write fast numeric
Web Page
March 25, 2014
Finite Volume schemes for the com-
of hands-on sessions.
algorithms for NVIDIA graphics proces-
www.hlrs.de/training/course-list
LRZ Building, University campus
pressible Euler equations. Hands-on
Day 1
sors (GPUs). Focus is on the basic
Garching near Munich.
sessions will manifest the content of
Intel tools: MPI tracing and Checking
usage of the language, the exploitation
compute/courses
Cray XE6/XC 30
Optimization Workshop
(PATC course)
in collaboration of Erlangen Regional
Iterative Linear Solvers and
Parallelization
Date and Location
the lectures. Participants will learn
Intel tools: OpenMP performance and
of the most important features of the
Contents
to implement the algorithms, but
correctness.
device (massive parallel computation,
This course is targeted at scientists
also to apply existing software and to
Day 2
shared memory, texture memory) and
March 17 - 20, 2014
Dates and Location
who wish to be introduced to pro-
interpret the solutions correctly.
Parallel I/O with MPI IO
efficient usage of the hardware to
Stuttgart, HLRS
March 24 - 28, 2014
gramming C/C++/Fortran with the
Methods and problems of paralleliza-
Performance analysis with Scalasca.
maximize performance. An overview
Stuttgart, HLRS
Eclipse C/C++ Development Tools
tion are discussed. This course is
Day 3
of the available development tools and
(CDT), or the Photran Plugin. Topics
based on a lecture and practical
Tuning I/O on LRZ's HPC systems.
the advanced features of the language
is given.
Contents
HLRS installed Hermit, a Cray XE6
September 15 - 19, 2014
covered include:
awarded with the "Landeslehrpreis
Portability of I/O: Binary files NetCDF
system with AMD Interlagos proces-
Garching, LRZ
s )NTRODUCTIONTO%CLIPSE)$%
Baden-Württemberg 2003" and or-
HDF5.
s )NTRODUCTIONTO#$4
ganized by HLRS, IAG, and University
Day 4
Web Page
of Kassel.
PGAS programming with coarray
www.hlrs.de/training/course-list
sors and a performance of 1 PFlop/s.
We strongly encourage you to port
Contents
s (ANDSONWITH#$4
your applications to the new architec-
The focus is on iterative and parallel
s 3HORTINTRODUCTIONANDDEMOOF ture as early as possible. To support
solvers, the parallel programming
such effort we invite current and
models MPI and OpenMP, and the
future users to participate in special
parallel middleware PETSc. Thereby,
Prerequisites
Cray XE6/XC30 Optimization Work-
different modern Krylov Subspace
Course participants should have basic
shops. With this course, we will give
Methods (CG, GMRES, BiCGSTAB ...)
all necessary information to move
as well as highly efficient precondi-
applications from the current NEC
tioning techniques are presented
SX-9, the Nehalem cluster, or other
in the context of real life applica-
Web Page
Date and Location
systems to Hermit. Hermit provides
tions. Hands-on sessions (in C and
http://www.lrz.de/services/
March 31 - April 03, 2014
Web Page
Many-core programming is a very
our users with a new level of perfor-
Fortran) will allow users to immedi-
compute/courses
LRZ Building, University campus
http://www.lrz.de/services/
dynamic research area. Many scien-
mance. To harvest this potential will
ately test and understand the basic
Garching near Munich.
compute/courses
tific applications have been ported to
require all our efforts. We are look-
constructs of iterative solvers, the
GPU architectures during the past
ing forward to working with our users
Message Passing Interface (MPI)
four years. We will give an introduction
Photran.
Fortran and Unified Parallel C.
Web Page
PGAS hands on session.
www.hlrs.de/training/course-list
GPU Programming
(PATC course)
Prerequisites
Good MPI and OpenMP knowledge
Date and Location
knowledge of the C and/or C++/
Advanced Topics in High
Performance Computing
as presented in the course "Parallel
April 07 - 09, 2014
Fortran programming languages.
(PATC course)
programming of High Performance
JSC, Forschungszentrum Jülich
Systems" (see above).
Contents
GCS – High Performance Computing
Scientific Visualization
Courses and Tutorials
will be presented, and interoperability
programming using examples of in-
of OpenACC directives with these
creasing complexity. After introducing
and with CUDA will be demonstrated.
Date and Location
become very popular, however pro-
Date and Location
MPI, OpenMP, performance optimi-
the basics the focus will be on
Through application case studies and
April 16 - 17, 2014
gramming GPGPUs using program-
May 05 - 06, 2014
zation, mathematical software, and
optimization and tuning of scientific
tutorials, users will gain direct expe-
Stuttgart, HLRS
ming languages like CUDA or OpenCL
JSC, Forschungszentrum Jülich
application software.
applications.
rience of using OpenACC directives in
This course is a PATC course (PRACE
realistic applications.
Contents
Beyond introducing the basics of
Contents
Web Page
Advanced Training Centres).
Users may also bring their own codes
This two day course is targeted at
GPGPU-porogramming, we mainly
Today's computers are commonly
http://www.fz-juelich.de/ias/jsc/
to discuss with Cray specialists or
researchers with basic knowledge
present OpenACC as an easier way
equipped with multicore processors
events/sc-may
begin porting.
in numerical simulation, who would
to program GPUs using OpenMP-like
and graphics processing units. To
Web Page
Open ACC Programming
Cray XK
(PATC course)
Particularly GPGPUs have recently
are: system architecture, usage
model, compilers, tools, monitoring,
is cumbersome and error-prone.
like to learn how to visualize their
pragmas. Recently Intel developed
make efficient use of these massively
Web Page
simulation results on the desktop but
their own Many Integrated Core
parallel compute resources advanced
Parallel I/O and Portable
Data Formats
www.hlrs.de/training/course-list
also in Augmented Reality and Virtual
(MIC) architecture which can be
knowledge of architecture and pro-
(PATC course)
Environments. It will start with a
programmed using standard paral-
gramming models is indispensable.
Unified Parallel C (UPC) and
Co-Array Fortran (CAF)
short overview of scientific visualiza-
lel programming techniques like
This course focuses on finding and
Date and Location
tion, following a hands-on introduc-
OpenMP and MPI. In the beginning
eliminating bottlenecks using profiling
May 21 - 23, 2014
(PATC course)
tion to 3D desktop visualization with
of 2013, the first production-level
and advanced programming tech-
JSC, Forschungszentrum Jülich
COVISE. On the second day, we will
cards named Intel Xeon Phi came on
niques, optimal usage of CPUs and
http://www.fz-juelich.de/ias/jsc/
events/gpu
computing since around 5 years.
Advanced GPU Programming
to CUDA, OpenCL, and multi-GPU
Date and Location
April 10 - 11, 2014
Date and Location
discuss how to build interactive 3D
the market. The course discusses
GPUs on a single node, and multi-GPU
Contents
Stuttgart, HLRS
April 14 - 15, 2014
Models for Virtual Environments and
various programming techniques for
programming across multiple nodes.
This course will introduce MPI parallel
Stuttgart, HLRS
how to set up an Augmented Reality
Intel Xeon Phi and includes hands-on
visualization.
session for both MIC and GPU pro-
Web Page
formats, such as HDF5 and NetCDF.
Contents
I/O and portable, self-describing data
This workshop will cover the pro-
Contents
gramming. The course is developed
http://www.fz-juelich.de/ias/jsc/
Participants should have experience
gramming environment of the Cray
Partitioned Global Address Space
Web Page
in collaboration with the Erlangen
events/advgpu
in parallel programming in general,
XK7 hybrid supercomputer, which
(PGAS) is a new model for parallel
www.hlrs.de/training/course-list
Regional Computing Centre (RRZE)
combines multicore CPUs with GPU
programming. Unified Parallel C (UPC)
accelerators. Attendees will learn
and Co-Array Fortran (CAF) are PGAS
extensions to C and Fortran. PGAS
Intel MIC&GPU
Programming Workshop
Each day is comprised of approxi-
about the directive-based OpenACC
programming model whose multi-
languages allow any processor to
(PATC course)
hours of hands-on sessions.
vendor support allows users to por-
directly address memory/data on any
tably develop applications for parallel
other processors.
Date and Location
Prerequisites
Date and Location
accelerated supercomputers.
Parallelism can be expressed more
April 28 - 30, 2014
Good working knowledge of at least
May 19 - 20, 2014
The workshop will also demonstrate
easily compared to library based
LRZ Building, University campus
one of the standard HPC languages:
JSC, Forschungszentrum Jülich
how to use the Cray Programming
approches as MPI. Hands-on sessions
Garching, near Munich.
Fortran 95, C or C++. Basic OpenMP
Environment tools to identify CPU
(in UPC and/or CAF) will allow users
application bottlenecks, facilitate the
to immediately test and understand
Contents
OpenACC porting, provide accelerated
the basic constructs of PGAS languages.
With the rapidly growing demand for
Web Page
supercomputers JUROPA and
computing power new accelerator
http://www.lrz.de/services/
JUQUEEN. Especially new users will
compute/courses
learn how to program and use these
performance feedback and to tune
within KONWIHR.
mately 5 hours of lectures and 2
and either C/C++ or Fortran in par-
Introduction to the
Programming and Usage
of the Supercomputer
Resources at Jülich
ticular. This course is a PATC course
(PRACE Advanced Training Centres).
Web Page
http://www.fz-juelich.de/ias/jsc/
and MPI knowledge useful.
the ported applications. The Cray
Web Page
based architectures have entered
scientific libraries for accelerators
www.hlrs.de/training/course-list
the world of high performance
Contents
This course gives an overview of the
systems efficiently. Topics discussed
events/parallelio
Authors
Momme Allalen
Dietmar Erwin
Ferdinand Jamitzky
Michael M. Resch
MOMMEALLALEn
EALLALEn
DERWin
fERDinAnDjAMiTzKy
RESCH
LRzDE
fz-JUELICHDE
LRzDE
HLRSDE
Fakherr F. Assaad
Jochen Fröhlich
Ivo Kabadshow
Helmut Satzger
ASSAAD PHySiKUni-WUERzBURGDE
jOCHEnfROEHLiCH TU-DRESDEnDE
iKABADSHOW
HELMUTSATzGER
Norbert
ert Attig
Andrea Gatto
Anupam Karmakar
Armin Seyfried
nATTiG
ASEyfRiED
fz-JUELICHDE
fz-JUELICHDE
AnDREAG MPA-GARCHInGMPGDE
AnUPAMKARMAKAR
Christian
tian Baczynski
Paul Gibbon
Tobias
T
Tob
ias Kempe
Wolfram Schmidt
BACzynSKi
SKi
PGiBBOn
TOBiASKEMPE
SCHMiDT
Uni-HDDE
fz-JUELICHDE
LRzDE
LRzDE
TU-DRESDEnDE
fz-JUELICHDE
ASTROPHySiK
Gurvan
n Bazin
Philipp Girichidis
Francisco Kitaura
Uni-GOETTinGEnDE
GURVAnBAZin
BAZin
PHILIPP
KiTAURA
Magnus Schwörer
GMAiLCOM
GiRICHiDiSCOM
AiPDE
Florian
n Berberich
Colin Glass
Ralf Klessen
MAGnUSSCHWOERER
fBERBERICH
RICH
GLASS
KLESSEn
GOOGLEMAiLCOM
fz-JUELICHDE
HLRSDE
Uni-HEiDELBERGDE
Christoph
toph Bernau
Simon Glover
Bastian Koller
Anton Stephan
CHRiSTOPHBERNAU
PHBERNAU
GLOVER
KOLLER
AnTOnSTEPHAn
LRzDE
Uni-HEiDELBERGDE
HLRSDE
DLRDE
Arndt Bode
Stefan Gottlöber
Carsten Kutzner
Sven Strohmer
ARnDTBODE
ODE
SGOTTLOEBER
CKUTznE
SSTROHMER
LRzDE
AiPDE
GWDGDE
fz-JUELICHDE
Maik Boltes
Jose Gracia
Thomas Lippert
Godehard Sutmann
MBOLTES
S
GRACiA
THLIPPERT
GSUTMAnn
fz-JUELICHDE
HLRSDE
fz-JUELICHDE
fz-JUELICHDE
David Brayford
Bärbel Große-W
Große-Wöhrmann
- öhrmann
-W
Daniel Mallmann
Jan Tr
T
Treibig
eibig
DAViDBRAyfORD
RAyfORD
WOEHRMAnn
DMALLMAnn
jAnTREiBiG
LRzDE
HLRSDE
fz-JUELICHDE
RRzEfAUDE
Matthias
hias Brehm
Carla Guillen Carias
Andreas Marek
Philipp Tr
T
Trisjono
isjono
MATTHiASBREHM
ASBREHM
CARLAGUILLEn
AMAREK
PTRISjOnO
LRzDE
LRzDE
RzGMPGDE
iTVRWTH-AACHEnDE
Dirk Brömmel
Nicolay Hammer
Gerald Mathias
Thomas Ullmann
DBROEMMEL
MMEL
niCOLAyHAMMER
GERALDMATHiAS
THOMASULLMAnn
fz-JUELICHDE
LRzDE
MPiBPCDE
Alexandru
ndru Calotoiu
Tim Hender
PHySiKUni-MUEnCHEnDE
Bernhard Vowinckel
ACALOTOiU
OiU
TiMHEnDER
Bernd Mohr
BERnHARDVOWinCKEL TU-DRESDEnDE
Alexeyy Cheptsov
Wolfram Hesse
BMOHR
Stefanie Walch
CHEPTSOV
OV
WOLfRAMHESSE
Thorsten Naab
WALCH
Regina Weigand
GRS-SiMDE
HLRSDE
CCfEACUK
LRzDE
Paul Clark
Steffen Heß
nAAB
PCLARK
SHESS
fz-JUELICHDE
MPA-GARCHinGMPGDE
MPA-GARCHinGMPGDE
Carmen Navarrete
WEIGAnD
Tanja Cleese
Torsten
T
To
rsten Hoefler
CARMEnnAVARRETE
Volker Weinberg
TAnjACLEES
EES
HTOR
Christoph Niethammer
VOLKERWEinBERG
Torsten
T
To
rsten Wilde
Uni-HEiDELBERGDE
SCAifRAUnHOfERDE
AiPDE
infETHzCH
Holgerr Dachsel
Martin Hohenadler
niETHAMMER
HDACHSEL
SEL
LRzDE
HLRSDE
HLRSDE
LRzDE
MARTinHOHEnADLER
Boris Orth
TORSTEnWiLDE
Ulrich Detert
PHySiKUni-WUERzBURGDE
BORTH
Felix Wolf
UDETERTT
Stefan Holl
Ludger Palm
fWOLf
Klaus Dolag
STHOLL
LUDGERPALM
Klaus Wolkersdorfer
KDOLAG
fz-JUELICHDE
fz-JUELICHDE
fz-JUELICHDE
fz-JUELICHDE
LRzDE
LRzDE
GRS-SiMDE
Frank Holzäpfel
Thomas Peters
KWOLKERSDORfER
Jan Frederik
rederik Engels
fRAnKHOLzAEPfEL
TPETERS
Richard Wünsch
MAiL
Valentina Huber
Rolf Rabenseifner
VHUBER
RABEnSEifnER
MPA-GARCHinGMPGDE
jfEnGELSDE
EnGELSDE
DLRDE
fz-JUELICHDE
)FYOUWOULDLIKETORECEIVEINSIDEREGULARLY
SENDANEMAILWITHYOURPOSTALADDRESSTOKLANK HLRSDE
7EBPAGEHTTPINSIDEHLRSDE
PHySiKUzHCH
RICHARD
fz-jUELiCHDE
WUnSCHCz
HLRSDE
©(,23