Editorial Editorial Publishers Editor Design 2 3 Contents Contents Contents News Applications Systems Centres Activities Projects Courses News News PRACE: Results of the 7th Regular Call PRACE Projects meet in Varna News News Interview with Prof. Dr.-Ing. Dr. h.c. Dr. h.c. Michael M. Resch News News Prof. Michael M. Resch is the Chairman of the Board of Directors of GCS since May 2013. He was a cofounder of GCS. Prof. Resch is currently the director of the High Performance Computing Center Stuttgart (HLRS), the director of the Information Service Center (IZUS) of the University of Stuttgart and the director of the Institute for High Performance Computing (IHR) of the University of Stuttgart. Applications Applications Simulating the Life Cycle of molecular Clouds 2 2 Numerical Method Applications Applications The typical Milky Way Disk 2 2 2 2 2 Applications Applications Disk Galaxies at different Gas Surface Densities: from low to high Redshift References [1] 2 2 Conclusions and Outlook Bigiel, F., Leroy, A., Walter, F., Brinks, E., de Blok, W. J. G., Madore, B., Thornley, M. D. [2] Brown, P. N., Byrne, G. D., Hindmarsh, A. C. [3] Chabrier, G. [4] Clark, P. C., Glover, S. C. O., Klessen, R. S. 2 [5] Evans, II, N. J. and et al. [6] Fryxell, B., et al. [7] 2 Glover, S. C. O., Mac Low, M.-M. [8] Górski, K. M., Hivon, E. 2 2 3 2 [9] Mac Low, M.-M., Klessen, R. S. 2 2 [10] Marinacci, F., Fraternali, F., Nipoti, C., Binney, J., Ciotti, L., Londrillo, P. 2 [11] Moster, B. P., Somerville, R. S., Maulbetsch, C., van den Bosch, F. C., Macciò, A. V., Naab, T., Oser, L. [12] Newman, S. F., et al. 2 [13] Tacconi, L., et al. 3 Applications Applications Gadget3: Numerical Simulation of Structure Formation in the Universe 3 Applications Applications References [1] Springel, V. [2] Springel, V., Yoshida, N., White, S.D.M. 2 2 Applications Applications Numerical Simulation of Correlated Electron Systems The Kane-Mele-Hubbard Model The Numerical Challenge 22 23 Applications Applications Results Impact of U for 2 =0 Applications Applications [11] Hohenadler, M., Meng, Z. Y., Lang, T. C., Wessel, S., Muramatsu, A., Assaad, F. F. [12] Berg, E., Metlitski, M.A., Sachdev, S. Acknowledgments [13] Rachel, S., Le Hur, K. [14] Assaad, F.F., Bercx, M., Hohenadler, M. [15] Herbut, I.F. Impact of U for >0 [16] Herbut, I.F., Juri i V., Vafek, O. [17] Yang, H.-Y., Albuquerque, A.F., Capponi, S., Läuchli, A.M., Schmidt, K.P. References [1] [18] Chang, C.-C., Scalettar, R.T. Kane, C.L., Mele, E.J. [19] Clark, B.K. [2] König, M., Wiedmann, S., Brüne, C., Roth, A., Buhmann, H., Molenkamp, L.W., Qi, X.-L., Zhang, S.-C. [3] Meng, Z.Y., Lang, T.C., Wessel, S., Assaad, F.F., Muramatsu, A. [20] Sorella, S., Tosatti, E. [21] Paiva, T., Scalettar, R.T., Zheng, W., Singh, R.R.P., Oitmaa, J. [22] Qi, X.-L., Zhang, S.-C. [4] Sorella, S., Otsuka, Y., Yunoki, S. [23] Ran, Y., Vishwanath, A., Lee, D.-H. [5] Assaad, F.F., Herbut, I.F. [6] Hohenadler, M., Assaad, F.F. [7] Outlook Uehlinger, T., Jotzu, G., Messer, M., Greif, D., Hofstetter, W., Bissbort, U., Esslinger, T. [8] Assaad, F.F., Evertz, H.G. [9] Assaad, F.F. [10] Hohenadler, M., Lang, T.C., Assaad, F.F. Applications Applications Highly-resolved numerical Simulations of bed-load Transport in a turbulent open Channel Flow Numerical Method Computational Setup Results Applications Applications References [1] Dietrich, W.E., Kirchner, J.W., Ikeda, H., Iseya, F. [2] Shvidchenko, A.B., Pender, G. [3] Kempe, T., Fröhlich, J., [4] Kempe, T., Fröhlich, J., [5] Vowinckel, B., Kempe, T., Fröhlich, J., Conclusions 2.5 1 0.8 Fix Ref FewPart LowSh Ref FewPart LowSh 2 1.5 Gm( x,0) Ruu 0.6 0.4 0.2 [7] Yalin, M.S., Ferreira da Silva, A.M., [8] Vowinckel, B., Kempe, T., Fröhlich, J., Nikora, V.I. 1 0.5 0 -0.2 [6] Shields, A. 0 2 4 6 rx / H 8 10 12 0 0 2 4 x 6 /H 8 10 12 [9] Vowinckel, B., Kempe, T., Fröhlich, J. Applications Applications How to fit the Local Universe into a Supercomputer? Recovering and Simulating Structures of the Local Universe 32 33 Applications Applications Acknowledgements References [1] Heß, S., Kitaura, F.-S., Gottlöber, S. [2] Kitaura, F.-S. [3] Kitaura, F.-S., Heß, S. Applications Applications A scalable hybrid DFT/PMM-MD Approach for accurately simulating Biomolecules on SuperMUC Applications Applications References [1] Senn, H.M., Thiel, W. [2] Schwörer, M., Breitenfeld, B., Tröster, P., Bauer, S., Lorenzen, K., Tavan, P., Mathias, G. [3] Lorenzen, K., Schwörer, M., Tröster, P., Mates, S., Tavan, P. [4] [5] Mathias, G., Baer, M.D. 2 Applications Applications Aircraft Wake Vortex Evolution during Approach and Landing With and without Plate Lines Applications Applications 3 0 3 Applications Applications References [1] Holzäpfel, F., Steen, M. [2] Holzäpfel, F., Gerz, T., Frech, M., Tafferner, A., Köpp, F., Smalikho, I., Rahm, S., Hahn, K.-U., Schwarz, C. [3] Misaka, T., Holzäpfel, F., Gerz, T. [4] Stephan, A., Holzäpfel, F., Misaka, T. Links Projects Projects FORTISSIMO Factories of the Future Resources, Technology, Infrastructure and Services for Simulation and Modelling Projects Projects References [1] Sawyer, P. [2] Schubert, J. [3] Core Project Partners Projects Projects Revisiting Dynamic Scheduling Techniques for HPC Infrastructures: The Approach of the DreamCloud Project [3] Project Partners References [1] [2] [3] [4] [1] [2] [4] Projects Projects SkaSim - Scalable HPC-Codes for molecular Simulation in the Chemical Industry Projects Projects POLCA: Programming Large Scale Heterogeneous Infrastructures Projects Projects The POLCA Approach Programming with POLCA Projects Projects What POLCA will provide Who is POLCA? Projects Projects A flexible Framework for Energy and Performance Analysis of highly parallel Applications in a Supercomputing Centre Background Software Projects Projects Summary References [1] [2] [3] Focht, E., Jeutter, A. [4] Guillen, C., Hesse, W., Brehm, M. [5] [6] Treibig, J., Hager, G., Wellein, G. 2 Focus of the Project Implementation [7] [8] [9] Projects Projects SIMOPEK – Simulation and Optimization of Data Center Energy Flows from Cooling Networks taking into Account HPC Operation Scenarios External Influences/Constraints Data Center: Reduce Total Cost of Ownership Pillar 1 Building Infrastructure Pillar 2 HPC System Hardware Pillar 3 HPC System Software System Scheduler Infrastructure Aware Resource Management & Scheduling U!lity Providers Neighboring Buildings Advanced Heat Reuse Technologies Pillar 4 HPC Applications Modeling, Simula!on & Op!miza!on Data Center Data Acquisi!on Monitor Building Management & Infrastructure Hardware Management System Management So"ware Performance Analysis Tools Infrastructure Monitoring System Hardware Monitoring System So"ware Monitoring Performance Monitoring Projects Projects References [1] Wilde, T., Auweter, A., Shoukourian, H. [2] Shoukourian, H., Wilde, T., Auweter, A. External Influences/Constraints [3] Data Center: Reduce Total Cost of Ownership Pillar 1 Building Infrastructure Pillar 3 HPC System Software Pillar 2 HPC System Hardware SIMOPEK Power Consump!on Modeling, Simula!on & Op!miza!on using MYNTS Facts and Figures FEPA Performance, Energy Modeling & Op!miza!on FEPA Data Collec!on SIMOPEK Data Collec!on using PowerDam V.2.0 Building Management & Infrastructure Hardware Management [4] System Scheduler Infrastructure Aware Resource Management & Scheduling U!lity Providers Neighboring Buildings SIMOPEK Advanced Absorp!on Cooling Pillar 4 HPC Applications System Management So"ware 2 Performance Analysis Tools 2 Infrastructure Monitoring System Hardware Monitoring System So"ware Monitoring Performance Monitoring Projects Projects The Catwalk Project – A quick Development Path for Performance Models Projects Projects References [1] Calotoiu, A., Hoefler, T., Poke, M., Wolf, F. [2] Geimer, M., Wolf, F., Wylie, B.J.N., Ábrahám, E., Becker, D., Mohr, B. [3] an Mey, D., Biersdorff, S., Bischof, C., Diethelm, K., Eschweiler, D., Gerndt, M., Knüpfer, A., Lorenz, D., Malony, A.D., Nagel, W.E., Oleynik, Y., Rössel, C., Saviankou, P., Schmidl, D., Shende, S.S., Wagner, M., Wesarg, B., Wolf, F. Projects Projects GROMEX – Unified Longrange Electrostatics and Flexible Ionization Usability & Scalability Background of the Project Towards realistic Simulations Projects Projects Project Partners References (1) Kabadshow, I., Dachsel, H. (4) Hess, B., Kutzner, C., van der Spoel, D., Lindahl, E. (5) Ullmann, R.T., Ullmann, G.M. (2) Dachsel, H. (3) Donnini, S., Tegeler, F., Groenhof, G., Grubmüller, H. 2 Projects Projects HOPSA – A big Jump forward in HPC System and Application Monitoring Integration among the HOPSA Performance Analysis Tools The HOPSA Performance Tool Workflow Projects Projects Conclusion Integration of System Data and Performance Analysis Tools Projects Projects [5] an Mey, D., Biersdorff, S., Bischof, C., Diethelm, K., Eschweiler, D., Gerndt, M., Knüpfer, A., Lorenz, D., Malony, A.D., Nagel, W.E., Oleynik, Y., Rössel, C., Saviankou, P., Schmidl, D., Shende, S.S., Wagner, M., Wesarg, B., Wolf, F. Russian Project Partners (HOPSA-RU) [6] Servat, H., Llort, G., Giménez, J., Labarta, J. EU Project Partners (HOPSA-EU) [7] References [1] Labarta, J., Girona, S., Pillet, V., Cortes, T., Gregoris, L. [2] Geimer, M., Wolf, F., Wylie, B.J.N., Abraham, E., Becker, D., Mohr, B. [8] Adinets, A.V., Bryzgalov, P.A., Vad, V., Voevodin, V., Zhumatiy, S.A., Nikitenko, D.A. [9] Mohr, B., Voevodin, V., Giménez, J., Hagersten, E., Knüpfer, A., Nikitenko, D.A., Nilsson, M., Servat, H., Shah, A., Winkler, F., Wolf, F., Zhujov, I. [3] Berg, E., Hagersten, E. [10] [4] Nagel, W., Weber, M., Hoppe, H.-C., Solchenbach, K. Systems Systems End of the HPC-FF Era Systems Systems 2 References [1] [2] [3] [4] Systems Systems JUROPA-3 - A Prototype for the Next-Generation HPC Cluster System Specifications Systems Systems First Experiences with the Intel MIC Architecture at LRZ Architectural Overview Intel MIC Architecture at LRZ Programming Models Number of cores Frequency of cores GDDR5 memory size Number of hardware threads SIMD vector registere Flops/cycle Theoretical peak performance L2 cache per core Systems Systems Benchmarks Acknowledgements References (1) (2) (3) Weinberg, V., (Editor) et al. Systems Systems The Extension of SuperMUC: Phase 2 SuperMUC – Phase 1 Users from 25 European Countries Phase 2 Innovative Water Cooling Financing References (1) (2) (3) Centres Centres Leibniz Supercomputing Centre of the Bavarian Academy of Sciences and Humanities (Leibniz-Rechenzentrum, LRZ) provides comprehensive services to scientific and academic communities by: Research in HPC is carried out in collaboration with the distributed, statewide Competence Network for Technical and Scientific High Performance Computing in Bavaria (KONWIHR). Contact: Leibniz Supercomputing Centre Prof. Dr. Arndt Bode Boltzmannstr. 1 85478 Garching near Munich Germany Phone +49 - 89 - 358 - 31- 80 00 [email protected] www.lrz.de Compute servers currently operated by LRZ are given in the following table System Size Peak Performance (TFlop/s) Purpose User Community Centres Centres First German National Center Based on a long tradition in supercomputing at University of Stuttgart, HLRS (Höchstleistungsrechenzentrum Stuttgart) was founded in 1995 as the first German federal Centre for High Performance Computing. HLRS serves researchers at universities and research laboratories in Europe and Germany and their external and industrial partners with high-end computing power for engineering and scientific applications. Service for Industry Service provisioning for industry is done together with T-Systems, T-Systems sfr, and Porsche in the public-private joint venture hww (Höchstleistungsrechner für Wissenschaft und Wirtschaft). Through this co-operation industry always has acces to the most recent HPC technology. Bundling Competencies In order to bundle service resources in the state of Baden-Württemberg HLRS has teamed up with the Steinbuch Center for Computing of the Karlsruhe Institute of Technology. This collaboration has been implemented in the non-profit organization SICOS BW GmbH. World Class Research As one of the largest research centers for HPC HLRS takes a leading role in research. Participation in the German national initiative of excellence makes HLRS an outstanding place in the field. Contact: Höchstleistungsrechenzentrum Stuttgart (HLRS) Universität Stuttgart Prof. Dr.-Ing. Dr. h.c. Dr. h.c. Michael M. Resch Nobelstraße 19 70569 Stuttgart Germany Phone +49 - 711- 685 - 8 72 69 [email protected] / www.hlrs.de Hermit Compute servers currently operated by HLRS Peak Performance User Community System Size (TFlop/s) Purpose Cray XE6 "Hermit" (Q4 2011) 3,552 dual socket nodes with 113,664 AMD Interlagos cores 1,045 Capability Computing European and German Research Organizations and Industry NEC Cluster (Laki, Laki2) heterogenous compunting platform of 2 independent clusters 23 TB memory 9988 cores 911 nodes 170 Laki: 120,5 TFlops Laki2: 47,2 TFlops German Universities, Research Institutes and Industry Centres Centres Supercomputer-oriented research and development in selected fields of physics and other natural sciences by research groups of competence in supercomputing applications. The Jülich Supercomputing Centre (JSC) at Forschungszentrum Jülich enables scientists and engineers to solve grand challenge problems of high complexity in science and engineering in collaborative infrastructures by means of supercomputing and Grid technologies. Provision of supercomputer resources of the highest performance class for projects in science, research and industry in the fields of modeling and computer simulation including their methods. The selection of the projects is performed by an international peer-review procedure implemented by the John von Neumann Institute for Computing (NIC), a joint foundation of Forschungszentrum Jülich, Deutsches Elektronen-Synchrotron DESY, and GSI Helmholtzzentrum für Schwerionenforschung. Implementation of strategic support infrastructures including communityoriented simulation laboratories and cross-sectional groups on mathematical methods and algorithms and parallel performance tools, enabling the effective usage of the supercomputer resources. Higher education for master and doctoral students in cooperation e.g. with the German Research School for Simulation Sciences. Contact: Jülich Supercomputing Centre (JSC) Forschungszentrum Jülich Prof. Dr. Dr. Thomas Lippert 52425 Jülich Germany Phone +49 - 24 61- 61- 64 02 [email protected] www.fz-juelich.de/jsc Compute servers currently operated by JSC Peak Performance System Size (TFlop/s) ® Purpose User Community Activities Activities CECAM Tutorials at JSC 2 CHANGES Workshop Activities Activities Laboratory Experiments on Crowd Dynamics Activities Activities JSC Guest Student Programme on Scientific Computing 2013 Activities Activities High-Q Club – The highest scaling Codes on JUQUEEN Terra-Neo Gysela waLBerla PEPC PMG+PFASST Reference [1] dynQCD Brömmel, D. Activities Activities Jülich Supercomputing Centre contributes to visionary “Human Brain Project” References [1] [2] [3] Activities Activities Traffic and Granular Flow Conference celebrates 10th Edition by returning to Jülich UNICORE Summit 2013 References [1] [2] Reference [1] Activities Activities 3D Show at the Pharma Forum: Simulation and Visualization of the Airflow in Cleanrooms Links s Activities Activities The 17th HLRS-NEC Workshop on Sustained Simulation Performance s Activities Activities ls1 mardyn - a Massively Parallel Molecular Simulation Code Scalability Neighbour Search Dynamic Load Balancing Partners s s s s s s Activities Activities GCS at ISC’13 – Review GCS Booth Highlights 2 s Two GCS HPC systems amongst Top Ten of TOP500 ISC’13 Gauss Award Winner Activities Activities Extreme Scaling Workshop at LRZ July 9-11, 2013: Running Real World Applications on more than 130,000 Cores on SuperMUC s"1#$ISLANDS s#)!/ISLANDS s6ERTEXISLANDS s0'ADGET88,ISLANDS s'2/-!#3ISLANDS s,!--03ISLANDS s.YXISLANDS Activities Activities sWAL"%RLAISLANDS s!0%3ISLANDS LRZ Extreme Scale Benchmark and Optimization Suite s%XA-,ISLANDS s)#/.ISLANDS 3 IBM MPI 1.3 + icc 12.1 Intel MPI 4.1 + icc 13.1 81,000 atoms aquaporin, PME, 2 fs 2 M atoms ribosome, PME, 4 fs 12 M atoms peptides, PME, 2 fs s3EIS3OLISLANDS Performance Results Activities Activities HLRS Scientific Tutorials and Workshop Report and Outlook OpenACC Programming for Parallel Accelerated Supercomputers – an alternative to CUDA from Cray perspective Cray XE6/ XC30 Optimization Workshops PRACE Advanced Training Centre Parallel Programming Workshop 2014 – Workshop Announcements Iterative Solvers and Parallelization Scientific Conferences and Workshops at HLRS 12th HLRS/hww Workshop on Scalable Global Parallel File Systems (March/April 2014) 8th ZIH+HLRS Parallel Tools Workshop (date and location not yet fixed) High Performance Computing in Science and Engineering - The 17th Results and Review Workshop of the HPC Center Stuttgart (October 2014) IDC International HPC User Forum (October 2014) Parallel Programming Workshops: Training in Parallel Programming and CFD Parallel Programming and Parallel Tools (TU Dresden, ZIH, February 24 - 27) Introduction to Computational Fluid Dynamics (HLRS, March 31 - April 4) ISC and SC Tutorials Iterative Linear Solvers and Parallelization (HLRS, March 24-28) Georg Hager, Gabriele Jost, Rolf Rabenseifner: Hybrid Parallel Programming with MPI & OpenMP. Tutorial 9 at the International Supercomputing Conference, ISC’13, Leipzig, June 16-20. 2013. Georg Hager, Jan Treibig, Gerhard Wellein: Node-Level Performance Engineering. Tutorial 2 at the International Supercomputing Conference, ISC’13, Leipzig, June 16-20. 2013. Rolf Rabenseifner, Georg Hager, Gabriele Jost: Hybrid MPI and OpenMP Parallel Programming. Half-day Tutorial at Super Computing 2013, SC13, Denver, Colorado, USA, November 17-22, 2013. Cray XE6/XC30 Optimization Workshops (HLRS, March 17 - 20) (PATC) GPU Programming using CUDA (HLRS, April 7 - 9) Introduction to Computational Fluid Dynamics Open ACC Programming for Parallel Accelerated Supercomputers (HLRS, April 10 - 11) (PATC) MPI & OpenMP Unified Parallel C (UPC) and Co-Array Fortran (CAF) (HLRS, April 14 - 15) (PATC) Scientific Visualisation (HLRS, April 16 - 17) Fortran for Scientific Computing Parallel Programming with MPI & OpenMP (TU Hamburg-Harburg, July 28 - 30) Iterative Linear Solvers and Parallelization (LRZ, Garching, September 15 - 19) Introduction to Computational Fluid Dynamics (ZIMT Siegen, September/October) Message Passing Interface (MPI) for Beginners (HLRS, October 6 - 7) (PATC) Shared Memory Parallelization with OpenMP (HLRS, October 8) (PATC) Advanced Topics in Parallel Programming (HLRS, October 9 - 10) (PATC) Parallel Programming with MPI & OpenMP (FZ Jülich, JSC, December 1 - 3) Training in Programming Languages at HLRS Fortran for Scientific Computing (Dec 2 - 6, 2013 and Mar 10 - 14, 2014) (PATC) URLs: Unified Parallel C (UPC) and CoArray Fortran (CAF) http://www.hlrs.de/events/ http://www.hlrs.de/training/course-list/ (PATC): This is a PRACE PATC course GCS – High Performance Computing Parallel Programming with MPI, OpenMP and PETSc Date and Location model, compilers, tools, monitoring, computational science, and also sessions will allow users to immediately Contents MPI, OpenMP, performance optimi- show how an awareness of the per- test and understand the language This course is targeted at scientists zation, mathematical software, and formance features of an application constructs. with little or no knowledge of the application software. may lead to notable reductions in Contents Fortran programming language, but Date and Location Web Page needing it for participation in projects February 24 - 27, 2014 Web Page s )NTRODUCTION www.hlrs.de/training/course-list using a Fortran code base, for devel- Dresden, ZIH http://www.fz-juelich.de/ias/jsc/ s 0RACTICALPERFORMANCEANALYSIS events/sc-nov s -ICROBENCHMARKSANDTHEMEMORY The focus is on programming models hierarchy s 4YPICALNODELEVELSOFTWAREOVER sessions (in C and Fortran) will allow Node-Level Performance Engineering users to immediately test and (PATC course) s %XAMPLEPROBLEMS MPI, OpenMP, and PETSc. Hands-on understand the basic constructs of Contents tools like debugger and syntax checker The focus is on programming models (PATC course) as well as handling of compilers and MPI, OpenMP, and PETSc. Hands-on libraries. The language is for the most sessions (in C and Fortran) will allow part treated at the level of the Fortran users to immediately test and - The 3D Jacobi solver February 03 - 05, 2014 95 standard; features from Fortran understand the basic constructs of JSC, Forschungszentrum Jülich 2003 are limited to improvements on the Message Passing Interface (MPI) - The Lattice-Boltzmann Method and the shared memory directives of December 03 - 04, 2013 - Sparse Matrix-Vector Multiplication OpenMP. Course language is English. LRZ Building, - Backprojection algorithm for CT This course is organized by JSC in University Campus Garching, collaboration with HLRS. Presented near Munich, Boltzmannstr. 1 Contents getting acquainted with additional Date and Location Date and Location by Dr. Rolf Rabenseifner, HLRS. opment of their own codes, and for Second JUQUEEN Porting and Tuning Workshop heads the Message Passing Interface (MPI) the elementary level. Advanced and the shared memory directives of Contents Fortran features like object-oriented OpenMP. The last day is dedicated to The Blue Gene/Q petaflop super- programming or coarrays will be cov- tools for debugging and performance s %NERGY0ARALLEL3CALABILITY computer JUQUEEN marks another ered in a follow-on course in autumn. analysis of parallel applications. This Between each module, there is time quantum leap in supercomputer To consolidate the lecture material, course is organized by ZIH in collabo- for Questions and Answers! performance at JSC. In order to use each day's approximately 4 hours of ration with HLRS. this tool efficiently, special efforts by lecture are complemented by 3 hours of hands-on sessions. reconstruction Web Page This course teaches performance http://www.fz-juelich.de/ias/jsc/ engineering approaches on the Web Page the users are necessary, though. The events/mpi compute node level. “Performance http://www.lrz.de/services/ aim of this hands-on workshop is to engineering” as we define it is more compute/courses support current users of JUQUEEN Prerequisites in porting their software, in analyzing Course participants should have ba- its performance, and in improving sic UNIX/Linux knowledge (login with developing a thorough understanding Fortran for Scientific Computing its efficiency. This course is a PATC secure shell, shell commands, basic of the interactions between software (PATC course) course (PRACE Advanced Training programming, vi or emacs editors). Introduction to the Programming and Usage of the Supercomputer Resources at Jülich Parallel Programming with MPI, OpenMP, and Tools power consumption: November 25 - 27, 2013 JSC, Forschungszentrum Jülich Courses and Tutorials than employing tools to identify hotspots and bottlenecks. It is about and hardware. This process must Web Page www.hlrs.de/training/course-list Parallel Programming of High Performance Systems Dates and Location March 10 - 14, 2014 Centres). Date and Location start at the core, socket, and node Dates and Location Web Page RRZE building, University campus November 28 - 29, 2013 level, where the code gets executed December 02 - 06, 2013 and JSC, Forschungszentrum Jülich that does the actual computational March 10 - 14, 2014 Web Page http://www.lrz.de/services/com- Erlangen, Martensstr. 1: Via video http://www.fz-juelich.de/ias/jsc/ pute/courses conference at LRZ if there is sufficient work. Once the architectural require- Stuttgart, HLRS events/juqueenpt14 interest. Programming with Fortran Contents Contents ments of a code are understood and This course gives an overview of correlated with performance mea- Contents the supercomputers JUROPA and surements, the potential benefit of This course is dedicated for scientists JUQUEEN. Especially new users will optimizations can often be predicted. and students to learn (sequential) Dates and Locations Erlangen Regional Computing Centre learn how to program and use these We introduce a “holistic” node-level programming scientific applications February 03 - 07, 2014 (RRZE) and LRZ, is targeted at stu- systems efficiently. Topics discussed performance engineering strategy, with Fortran. The course teaches LRZ Building, University campus dents and scientists with interest in are: system architecture, usage apply it to different algorithms from newest Fortran standards. Hands-on Garching near Munich. programming modern HPC hardware, This course, a collaboration of GCS – High Performance Computing specifically the large scale parallel on these opportunities. From Monday and the shared memory directives of computing systems available in Munich, to Wednesday, specialists from Cray OpenMP. This course is organized by Jülich and Stuttgart. will support you in your effort porting University of Kassel, HLRS, and IAG. and optimizing your application on our Courses and Tutorials Introduction to Computational Fluids Dynamics Contents In this add-on course to the parallel GPU Programming using CUDA programming course special topics Date and Location are treated in more depth, in Date and Location Each day is comprised of approxi- Cray XE6. On Thursday, Georg Hager Web Page March 31 - April 04, 2014 particular performance analysis, I/O April 07 - 09, 2014 mately 4 hours of lectures and 3 and Jan Treibig from RRZE will present www.hlrs.de/training/course-list Stuttgart, HLRS and PGAS concepts. It is provided Stuttgart, HLRS hours of hands-on sessions. detailed information on optimizing Eclipse: C/C++/Fortran programming Contents Computing Centre (RRZE) and LRZ Contents Numerical methods to solve the within KONWIHR. The course provides an introduction equations of Fluid Dynamics are pre- Each day is comprised of approxima- to the programming language CUDA, codes on the multicore AMD Interlagos Web Page processor. Course language is English http://www.lrz.de/services/ (if required). Date and Location sented. The main focus is on explicit tely 5 hours of lectures and 2 hours which is used to write fast numeric Web Page March 25, 2014 Finite Volume schemes for the com- of hands-on sessions. algorithms for NVIDIA graphics proces- www.hlrs.de/training/course-list LRZ Building, University campus pressible Euler equations. Hands-on Day 1 sors (GPUs). Focus is on the basic Garching near Munich. sessions will manifest the content of Intel tools: MPI tracing and Checking usage of the language, the exploitation compute/courses Cray XE6/XC 30 Optimization Workshop (PATC course) in collaboration of Erlangen Regional Iterative Linear Solvers and Parallelization Date and Location the lectures. Participants will learn Intel tools: OpenMP performance and of the most important features of the Contents to implement the algorithms, but correctness. device (massive parallel computation, This course is targeted at scientists also to apply existing software and to Day 2 shared memory, texture memory) and March 17 - 20, 2014 Dates and Location who wish to be introduced to pro- interpret the solutions correctly. Parallel I/O with MPI IO efficient usage of the hardware to Stuttgart, HLRS March 24 - 28, 2014 gramming C/C++/Fortran with the Methods and problems of paralleliza- Performance analysis with Scalasca. maximize performance. An overview Stuttgart, HLRS Eclipse C/C++ Development Tools tion are discussed. This course is Day 3 of the available development tools and (CDT), or the Photran Plugin. Topics based on a lecture and practical Tuning I/O on LRZ's HPC systems. the advanced features of the language is given. Contents HLRS installed Hermit, a Cray XE6 September 15 - 19, 2014 covered include: awarded with the "Landeslehrpreis Portability of I/O: Binary files NetCDF system with AMD Interlagos proces- Garching, LRZ s )NTRODUCTIONTO%CLIPSE)$% Baden-Württemberg 2003" and or- HDF5. s )NTRODUCTIONTO#$4 ganized by HLRS, IAG, and University Day 4 Web Page of Kassel. PGAS programming with coarray www.hlrs.de/training/course-list sors and a performance of 1 PFlop/s. We strongly encourage you to port Contents s (ANDSONWITH#$4 your applications to the new architec- The focus is on iterative and parallel s 3HORTINTRODUCTIONANDDEMOOF ture as early as possible. To support solvers, the parallel programming such effort we invite current and models MPI and OpenMP, and the future users to participate in special parallel middleware PETSc. Thereby, Prerequisites Cray XE6/XC30 Optimization Work- different modern Krylov Subspace Course participants should have basic shops. With this course, we will give Methods (CG, GMRES, BiCGSTAB ...) all necessary information to move as well as highly efficient precondi- applications from the current NEC tioning techniques are presented SX-9, the Nehalem cluster, or other in the context of real life applica- Web Page Date and Location systems to Hermit. Hermit provides tions. Hands-on sessions (in C and http://www.lrz.de/services/ March 31 - April 03, 2014 Web Page Many-core programming is a very our users with a new level of perfor- Fortran) will allow users to immedi- compute/courses LRZ Building, University campus http://www.lrz.de/services/ dynamic research area. Many scien- mance. To harvest this potential will ately test and understand the basic Garching near Munich. compute/courses tific applications have been ported to require all our efforts. We are look- constructs of iterative solvers, the GPU architectures during the past ing forward to working with our users Message Passing Interface (MPI) four years. We will give an introduction Photran. Fortran and Unified Parallel C. Web Page PGAS hands on session. www.hlrs.de/training/course-list GPU Programming (PATC course) Prerequisites Good MPI and OpenMP knowledge Date and Location knowledge of the C and/or C++/ Advanced Topics in High Performance Computing as presented in the course "Parallel April 07 - 09, 2014 Fortran programming languages. (PATC course) programming of High Performance JSC, Forschungszentrum Jülich Systems" (see above). Contents GCS – High Performance Computing Scientific Visualization Courses and Tutorials will be presented, and interoperability programming using examples of in- of OpenACC directives with these creasing complexity. After introducing and with CUDA will be demonstrated. Date and Location become very popular, however pro- Date and Location MPI, OpenMP, performance optimi- the basics the focus will be on Through application case studies and April 16 - 17, 2014 gramming GPGPUs using program- May 05 - 06, 2014 zation, mathematical software, and optimization and tuning of scientific tutorials, users will gain direct expe- Stuttgart, HLRS ming languages like CUDA or OpenCL JSC, Forschungszentrum Jülich application software. applications. rience of using OpenACC directives in This course is a PATC course (PRACE realistic applications. Contents Beyond introducing the basics of Contents Web Page Advanced Training Centres). Users may also bring their own codes This two day course is targeted at GPGPU-porogramming, we mainly Today's computers are commonly http://www.fz-juelich.de/ias/jsc/ to discuss with Cray specialists or researchers with basic knowledge present OpenACC as an easier way equipped with multicore processors events/sc-may begin porting. in numerical simulation, who would to program GPUs using OpenMP-like and graphics processing units. To Web Page Open ACC Programming Cray XK (PATC course) Particularly GPGPUs have recently are: system architecture, usage model, compilers, tools, monitoring, is cumbersome and error-prone. like to learn how to visualize their pragmas. Recently Intel developed make efficient use of these massively Web Page simulation results on the desktop but their own Many Integrated Core parallel compute resources advanced Parallel I/O and Portable Data Formats www.hlrs.de/training/course-list also in Augmented Reality and Virtual (MIC) architecture which can be knowledge of architecture and pro- (PATC course) Environments. It will start with a programmed using standard paral- gramming models is indispensable. Unified Parallel C (UPC) and Co-Array Fortran (CAF) short overview of scientific visualiza- lel programming techniques like This course focuses on finding and Date and Location tion, following a hands-on introduc- OpenMP and MPI. In the beginning eliminating bottlenecks using profiling May 21 - 23, 2014 (PATC course) tion to 3D desktop visualization with of 2013, the first production-level and advanced programming tech- JSC, Forschungszentrum Jülich COVISE. On the second day, we will cards named Intel Xeon Phi came on niques, optimal usage of CPUs and http://www.fz-juelich.de/ias/jsc/ events/gpu computing since around 5 years. Advanced GPU Programming to CUDA, OpenCL, and multi-GPU Date and Location April 10 - 11, 2014 Date and Location discuss how to build interactive 3D the market. The course discusses GPUs on a single node, and multi-GPU Contents Stuttgart, HLRS April 14 - 15, 2014 Models for Virtual Environments and various programming techniques for programming across multiple nodes. This course will introduce MPI parallel Stuttgart, HLRS how to set up an Augmented Reality Intel Xeon Phi and includes hands-on visualization. session for both MIC and GPU pro- Web Page formats, such as HDF5 and NetCDF. Contents I/O and portable, self-describing data This workshop will cover the pro- Contents gramming. The course is developed http://www.fz-juelich.de/ias/jsc/ Participants should have experience gramming environment of the Cray Partitioned Global Address Space Web Page in collaboration with the Erlangen events/advgpu in parallel programming in general, XK7 hybrid supercomputer, which (PGAS) is a new model for parallel www.hlrs.de/training/course-list Regional Computing Centre (RRZE) combines multicore CPUs with GPU programming. Unified Parallel C (UPC) accelerators. Attendees will learn and Co-Array Fortran (CAF) are PGAS extensions to C and Fortran. PGAS Intel MIC&GPU Programming Workshop Each day is comprised of approxi- about the directive-based OpenACC programming model whose multi- languages allow any processor to (PATC course) hours of hands-on sessions. vendor support allows users to por- directly address memory/data on any tably develop applications for parallel other processors. Date and Location Prerequisites Date and Location accelerated supercomputers. Parallelism can be expressed more April 28 - 30, 2014 Good working knowledge of at least May 19 - 20, 2014 The workshop will also demonstrate easily compared to library based LRZ Building, University campus one of the standard HPC languages: JSC, Forschungszentrum Jülich how to use the Cray Programming approches as MPI. Hands-on sessions Garching, near Munich. Fortran 95, C or C++. Basic OpenMP Environment tools to identify CPU (in UPC and/or CAF) will allow users application bottlenecks, facilitate the to immediately test and understand Contents OpenACC porting, provide accelerated the basic constructs of PGAS languages. With the rapidly growing demand for Web Page supercomputers JUROPA and computing power new accelerator http://www.lrz.de/services/ JUQUEEN. Especially new users will compute/courses learn how to program and use these performance feedback and to tune within KONWIHR. mately 5 hours of lectures and 2 and either C/C++ or Fortran in par- Introduction to the Programming and Usage of the Supercomputer Resources at Jülich ticular. This course is a PATC course (PRACE Advanced Training Centres). Web Page http://www.fz-juelich.de/ias/jsc/ and MPI knowledge useful. the ported applications. The Cray Web Page based architectures have entered scientific libraries for accelerators www.hlrs.de/training/course-list the world of high performance Contents This course gives an overview of the systems efficiently. Topics discussed events/parallelio Authors Momme Allalen Dietmar Erwin Ferdinand Jamitzky Michael M. Resch MOMMEALLALEn EALLALEn DERWin fERDinAnDjAMiTzKy RESCH LRzDE fz-JUELICHDE LRzDE HLRSDE Fakherr F. Assaad Jochen Fröhlich Ivo Kabadshow Helmut Satzger ASSAAD PHySiKUni-WUERzBURGDE jOCHEnfROEHLiCH TU-DRESDEnDE iKABADSHOW HELMUTSATzGER Norbert ert Attig Andrea Gatto Anupam Karmakar Armin Seyfried nATTiG ASEyfRiED fz-JUELICHDE fz-JUELICHDE AnDREAG MPA-GARCHInGMPGDE AnUPAMKARMAKAR Christian tian Baczynski Paul Gibbon Tobias T Tob ias Kempe Wolfram Schmidt BACzynSKi SKi PGiBBOn TOBiASKEMPE SCHMiDT Uni-HDDE fz-JUELICHDE LRzDE LRzDE TU-DRESDEnDE fz-JUELICHDE ASTROPHySiK Gurvan n Bazin Philipp Girichidis Francisco Kitaura Uni-GOETTinGEnDE GURVAnBAZin BAZin PHILIPP KiTAURA Magnus Schwörer GMAiLCOM GiRICHiDiSCOM AiPDE Florian n Berberich Colin Glass Ralf Klessen MAGnUSSCHWOERER fBERBERICH RICH GLASS KLESSEn GOOGLEMAiLCOM fz-JUELICHDE HLRSDE Uni-HEiDELBERGDE Christoph toph Bernau Simon Glover Bastian Koller Anton Stephan CHRiSTOPHBERNAU PHBERNAU GLOVER KOLLER AnTOnSTEPHAn LRzDE Uni-HEiDELBERGDE HLRSDE DLRDE Arndt Bode Stefan Gottlöber Carsten Kutzner Sven Strohmer ARnDTBODE ODE SGOTTLOEBER CKUTznE SSTROHMER LRzDE AiPDE GWDGDE fz-JUELICHDE Maik Boltes Jose Gracia Thomas Lippert Godehard Sutmann MBOLTES S GRACiA THLIPPERT GSUTMAnn fz-JUELICHDE HLRSDE fz-JUELICHDE fz-JUELICHDE David Brayford Bärbel Große-W Große-Wöhrmann - öhrmann -W Daniel Mallmann Jan Tr T Treibig eibig DAViDBRAyfORD RAyfORD WOEHRMAnn DMALLMAnn jAnTREiBiG LRzDE HLRSDE fz-JUELICHDE RRzEfAUDE Matthias hias Brehm Carla Guillen Carias Andreas Marek Philipp Tr T Trisjono isjono MATTHiASBREHM ASBREHM CARLAGUILLEn AMAREK PTRISjOnO LRzDE LRzDE RzGMPGDE iTVRWTH-AACHEnDE Dirk Brömmel Nicolay Hammer Gerald Mathias Thomas Ullmann DBROEMMEL MMEL niCOLAyHAMMER GERALDMATHiAS THOMASULLMAnn fz-JUELICHDE LRzDE MPiBPCDE Alexandru ndru Calotoiu Tim Hender PHySiKUni-MUEnCHEnDE Bernhard Vowinckel ACALOTOiU OiU TiMHEnDER Bernd Mohr BERnHARDVOWinCKEL TU-DRESDEnDE Alexeyy Cheptsov Wolfram Hesse BMOHR Stefanie Walch CHEPTSOV OV WOLfRAMHESSE Thorsten Naab WALCH Regina Weigand GRS-SiMDE HLRSDE CCfEACUK LRzDE Paul Clark Steffen Heß nAAB PCLARK SHESS fz-JUELICHDE MPA-GARCHinGMPGDE MPA-GARCHinGMPGDE Carmen Navarrete WEIGAnD Tanja Cleese Torsten T To rsten Hoefler CARMEnnAVARRETE Volker Weinberg TAnjACLEES EES HTOR Christoph Niethammer VOLKERWEinBERG Torsten T To rsten Wilde Uni-HEiDELBERGDE SCAifRAUnHOfERDE AiPDE infETHzCH Holgerr Dachsel Martin Hohenadler niETHAMMER HDACHSEL SEL LRzDE HLRSDE HLRSDE LRzDE MARTinHOHEnADLER Boris Orth TORSTEnWiLDE Ulrich Detert PHySiKUni-WUERzBURGDE BORTH Felix Wolf UDETERTT Stefan Holl Ludger Palm fWOLf Klaus Dolag STHOLL LUDGERPALM Klaus Wolkersdorfer KDOLAG fz-JUELICHDE fz-JUELICHDE fz-JUELICHDE fz-JUELICHDE LRzDE LRzDE GRS-SiMDE Frank Holzäpfel Thomas Peters KWOLKERSDORfER Jan Frederik rederik Engels fRAnKHOLzAEPfEL TPETERS Richard Wünsch MAiL Valentina Huber Rolf Rabenseifner VHUBER RABEnSEifnER MPA-GARCHinGMPGDE jfEnGELSDE EnGELSDE DLRDE fz-JUELICHDE )FYOUWOULDLIKETORECEIVEINSIDEREGULARLY SENDANEMAILWITHYOURPOSTALADDRESSTOKLANK HLRSDE 7EBPAGEHTTPINSIDEHLRSDE PHySiKUzHCH RICHARD fz-jUELiCHDE WUnSCHCz HLRSDE ©(,23
© Copyright 2024 ExpyDoc