+ All Categories
Home > Documents > Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 ·...

Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 ·...

Date post: 20-May-2020
Category:
Upload: others
View: 1 times
Download: 0 times
Share this document with a friend
30
Center for Computational Sciences, Univ. of Tsukuba Taisuke Boku Deputy Director, Center for Computational Sciences University of Tsukuba (with courtesy of JCAHPC members) Oakforest-PACS (OFP) : Japan’s Fastest Supercomputer 2017/04/05 1 1st F-J-G Workshop, Tokyo
Transcript
Page 1: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Taisuke BokuDeputy Director, Center for Computational Sciences

University of Tsukuba

(with courtesy of JCAHPC members)

Oakforest-PACS (OFP) :Japan’s Fastest Supercomputer

2017/04/051 1st F-J-G Workshop, Tokyo

Page 2: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Outlinen Supercomputer deployment plan in Japann What is JCAHPC ?n Supercomputer procurement in JCAHPCn Oakforest-PACS Systemn Records and evaluation so farn Summary

2017/04/051st F-J-G Workshop, Tokyo

2

Page 3: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Towards Exascale Computing

2017/04/05

1st F-J-G Workshop, Tokyo3

1

10

100

1000 Post K Computer

T2K

PF

2008 2010 2012 2014 2016 2018 2020

U. of TsukubaU. of TokyoKyoto U.

Riken AICS

FutureExascale

Tokyo Tech.TSUBAME2.0

Tier-1 and tier-2 supercomputers form HPCI and move forward to Exascale computing like two wheels

JCAHPC(U. Tsukuba and U. Tokyo)

OFP

Page 4: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Fiscal Year 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025

Hokkaido

Tohoku

Tsukuba

Tokyo

Tokyo Tech.

Nagoya

Kyoto

Osaka

Kyushu

Deployment plan of 9 supercomputing center (Feb. 2017)

HITACHI SR16000/M1(172TF, 22TB)Cloud System BS2000 (44TF, 14TB)Data Science Cloud / Storage HA8000 / WOS7000

(10TF, 1.96PB)

50+ PF (FAC) 3.5MW 200+ PF (FAC) 6.5MW

Fujitsu FX10 (1PFlops, 150TiB, 408 TB/s), Hitachi SR16K/M1 (54.9 TF, 10.9 TiB, 28.7 TB/s)

NEC SX-9他(60TF)

~30PF, ~30PB/s Mem BW (CFL-D/CFL-M) ~3MW

SX-ACE(707TF,160TB, 655TB/s)LX406e(31TF), Storage(4PB), 3D Vis, 2MW

100+ PF (FAC/UCC+CFL-M)up to 4MW

50+ PF (FAC/UCC + CFL-M)FX10(90TF) Fujitsu FX100 (2.9PF, 81 TiB)CX400(470TF) Fujitsu CX400 (774TF, 71TiB)SGI UV2000 (24TF, 20TiB) 2MW in total up to 4MW

Power consumption indicates maximum of power supply (includes cooling facility)

Cray: XE6 + GB8K +XC30 (983TF) 5 PF(FAC/TPF)

1.5 MWCray XC30 (584TF)

50-100+ PF(FAC/TPF + UCC) 1.8-2.4 MW

3.2 PF (UCC + CFL/M) 0.96MW 30 PF (UCC + CFL-M) 2MW0.3 PF (Cloud) 0.36MW

15-20 PF (UCC/TPF)HA8000 (712TF, 242 TB)SR16000 (8.2TF, 6TB) 100-150 PF

(FAC/TPF + UCC/TPF)FX10 (90.8TFLOPS)

3MWFX10 (272.4TF, 36 TB)CX400 (966.2 TF, 183TB)

2.0MW 2.6MW

HA-PACS (1166 TF)

100+ PF 4.5MW(UCC + TPF)

TSUBAME 4.0 (100+ PF, >10PB/s, ~2.0MW)

TSUBAME 2.5 (5.7 PF, 110+ TB, 1160 TB/s), 1.4MW

TSUBAME 2.5 (3~4 PF, extended)

25.6 PB/s, 50-100Pflop/s,1.5-2.0MW

3.2PB/s, 5-10Pflop/s, 1.0-1.5MW (CFL-M)NEC SX-ACE NEC Express5800(423TF) (22.4TF) 0.7-1PF (UCC)

COMA (PACS-IX) (1001 TF)

Reedbush 1.8〜1.9 PF 0.7 MWOakforest-PACS 25 PF (UCC + TPF) 4.2 MW

PACS-X 10PF (TPF) 2MW

2017/04/05 1st F-J-G Workshop, Tokyo4

TSUBAME3.012PF,UCC/TPF2.0MW

K Computer (10PF, 13MW) Post-K Computer (??. ??)????

Page 5: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Univ. Tsukuba648 nodes (95.4TF) / 20TBLinpack Result:

Rpeak= 92.0TF (625 nodes)Rmax = 76.5TF

Univ. Tokyo952 nodes (140.1TF) / 31TBLinpack Result:

Rpeak= 113.1TF (512+256 nodes)Rmax = 83.0TF

Kyoto Univ.416 nodes (61.2TF) / 13TBLinpack Result:

Rpeak= 61.2TF (416 nodes)Rmax = 50.5TF

n Same timing of procurement for next generation supercomputers in three universities

n Academic leadership for computational science/engineering in research/education/grid-use on same platform

n Open hardware architecture withcommodity devices & technologies.

n Open software stack with open-source middleware & tools.

n Open to user’s needs not only inFP & HPC field but also INT world.

2017/04/051st F-J-G Workshop, Tokyo

5

T2K Open Supercomputer Systems

Page 6: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

From T2K to Post-T2K

n Effect of T2K Alliancen Three supercomputers are introduced at the same time, sharing wide

knowledge for system construction and commodity technology, followed by academic research collaboration among these players

n After T2K, three universities had different time of new system procurementn Kyotop U.: four year period of procurementn U. Tsukuba: accelerated computingn U. Tokyo: T2K + Fujitsu FX10 and other systems

n Post-T2K (with two ”Ts”)n in 2013, U. Tsukuba and U. Tokyo collaborated again for new

supercomputer procurement in much more tight framework

2017/04/051st F-J-G Workshop, Tokyo6

Page 7: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

JCAHPC

n Joint Center for Advanced High Performance Computing (http://jcahpc.jp)

n Very tight collaboration for “post-T2K” with two universitiesn For main supercomputer resources, uniform specification to

single shared systemn Each university is financially responsible to introduce the

machine and its operation-> unified procurement toward single system with largest scale in Japan

n To manage everything smoothly, a joint organization was established-> JCAHPC

2017/04/051st F-J-G Workshop, Tokyo7

Page 8: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Procurement Policies of JCAHPCn based on the spirit of T2K, introducing open advanced technology

n massively parallel PC clustern advanced processor for HPCn easy to use and efficient interconnectionn large scale shared file system flatly shared by all nodes

n joint procurement by two universitiesn the largest class of budget as national universities’ supercomputer in Japann the largest system scale as PC cluster in Japann no accelerator to support wide variety of users and application fields

-> not chasing absolute peak performance and inheriting traditional application codes (basically)

n goodness of single systemn scale-merit by merging budget -> largest in Japann ultra large scale single job execution at special occasion such as “Gordon Bell Prize

Challenge”

2017/04/051st F-J-G Workshop, Tokyo8

⇒ Oakforest-PACS (OFP)

Page 9: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

2017/04/05

1st F-J-G Workshop, Tokyo9

Oakforest-PACS (OFP)

• 25 PFLOPS peak• 8208 KNL CPUs• FBB Fat-Tree by

OmniPath• HPL 13.55 PFLOPS

#1 in Japan#6 in World

• HPCG #3 in World• Green500 #6 in World

• Full operation started Dec. 2016

• Official Program started on April 2017

U. Tokyo convention U. Tsukuba convention ⇒ Don’t call it just “Oakforest” !“OFP” is much better

Page 10: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Computation node & chassis

2017/04/05

1st F-J-G Workshop, Tokyo10

Computation node (Fujitsu next generation PRIMERGY)with single chip Intel Xeon Phi (Knights Landing, 3+TFLOPS)and Intel Omni-Path Architecture card (100Gbps)

Chassis with 8 nodes, 2U size

Water cooling wheel & pipe

Page 11: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Water cooling pipes and IME (burst buffer)

2017/04/05

1st F-J-G Workshop, Tokyo11

Page 12: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Specification of Oakforest-PACS

2017/04/051st F-J-G Workshop, Tokyo12

Total peak performance 25 PFLOPS

Total number of compute nodes 8,208

Compute node

Product Fujitsu Next-generation PRIMERGY server for HPC (under development)

Processor Intel® Xeon Phi™ (Knights Landing)Xeon Phi 7250 (1.4GHz TDP) with 68 cores

Memory High BW 16 GB, > 400 GB/sec (MCDRAM, effective rate)

Low BW 96 GB, 115.2 GB/sec (DDR4-2400 x 6ch, peak rate)

Inter-connect

Product Intel® Omni-Path Architecture

Link speed 100 Gbps

Topology Fat-tree with full-bisection bandwidth

Login node

Product Fujitsu PRIMERGY RX2530 M2 server

# of servers 20

Processor Intel Xeon E5-2690v4 (2.6 GHz 14 core x 2 socket)

Memory 256 GB, 153 GB/sec (DDR4-2400 x 4ch x 2 socket)

Page 13: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Specification of Oakforest-PACS (I/O)

2017/04/05

1st F-J-G Workshop, Tokyo13

Parallel File System

Type Lustre File System

Total Capacity 26.2 PB

Metadata

Product DataDirect Networks MDS server + SFA7700X

# of MDS 4 servers x 3 set

MDT 7.7 TB (SAS SSD) x 3 set

Object storage

Product DataDirect Networks SFA14KE

# of OSS (Nodes)

10 (20)

Aggregate BW ~500 GB/sec

Fast File Cache System

Type Burst Buffer, Infinite Memory Engine (by DDN)

Total capacity 940 TB (NVMe SSD, including parity data by erasure coding)

Product DataDirect Networks IME14K

# of servers (Nodes) 25 (50)

Aggregate BW ~1,560 GB/sec

Page 14: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Full bisection bandwidth Fat-tree by Intel® Omni-Path Architecture

2017/04/051st F-J-G Workshop, Tokyo14

12 of768 port Director Switch(Source by Intel)

362 of48 port Edge Switch

2 2

241 4825 7249

Uplink: 24

Downlink: 24

. . . . . . . . . Compute Nodes 8208

Login Nodes 20

Parallel FS 64

IME 300

Mgmt, etc. 8

Total 8600

Firstly, to reduce switches&cables, we considered :• All the nodes into subgroups are connected with FBB Fat-tree• Subgroups are connected with each other with >20% of FBBBut, HW quantity is not so different from globally FBB, and globally FBB ispreferredfor flexible job management.

Page 15: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Facility of Oakforest-PACS system

2017/04/051st F-J-G Workshop, Tokyo15

Power consumption 4.2 MW (including cooling)

# of racks 102

Cooling system

Compute Node

Type Warm-water coolingDirect cooling (CPU)Rear door cooling (except CPU)

Facility Cooling tower & Chiller

Others Type Air cooling

Facility PAC

Page 16: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Software of Oakforest-PACS

2017/04/051st F-J-G Workshop, Tokyo16

Compute node Login node

OS CentOS 7, McKernel Red Hat Enterprise Linux 7Compiler gcc, Intel compiler (C, C++, Fortran)MPI Intel MPI, MVAPICH2Library Intel MKL

LAPACK, FFTW, SuperLU, PETSc, METIS, Scotch, ScaLAPACK, GNU Scientific Library, NetCDF, Parallel netCDF, Xabclib, ppOpen-HPC, ppOpen-AT, MassiveThreads

Application mpijava, XcalableMP, OpenFOAM, ABINIT-MP, PHASE system, FrontFlow/blue, FrontISTR, REVOCAP, OpenMX, xTAPP, AkaiKKR, MODYLAS, ALPS, feram, GROMACS, BLAST, R packages, Bioconductor, BioPerl, BioRuby

Distributed FS Globus Toolkit, GfarmJob Scheduler Fujitsu Technical Computing SuiteDebugger Allinea DDTProfiler Intel VTune Amplifier, Trace Analyzer & Collector

Page 17: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

TOP500 list on Nov. 2016

2017/04/05

1st F-J-G Workshop, Tokyo 17

# Machine Architecture Country Rmax (TFLOPS) Rpeak (TFLOPS) MFLOPS/W

1 TaihuLight, NSCW MPP (Sunway, SW26010) China 93,014.6 125,435.9 6051.3

2 Tianhe-2 (MilkyWay-2), NSCG

Cluster (NUDT, CPU + KNC) China 33,862.7 54,902.4 1901.5

3 Titan, ORNL MPP (Cray, XK7: CPU + GPU) United States 17,590.0 27,112.5 2142.8

4 Sequoia, LLNL MPP (IBM, BlueGene/Q) United States 17,173.2 20,132.7 2176.6

5 Cori, NERSC-LBNL MPP (Cray, XC40: KNL) United States 14,014.7 27,880.7 ???

6 Oakforest-PACS, JCAHPC Cluster (Fujitsu, KNL) Japan 13,554.6 25,004.9 4985.1

7 K Computer, RIKEN AICS MPP (Fujitsu) Japan 10,510.0 11,280.4 830.2

8 Piz Daint, CSCS MPP (Cray, XC50:CPU + GPU) Switzerland 9,779.0 15,988.0 7453.5

9 Mira, ANL MPP (IBM, BlueGene/Q) United States 8,586.6 10,066.3 2176.6

10 Trinity, NNSA/LABNL/SNL

MPP (Cray, XC40: MIC) United States 8,100.9 11,078.9 1913.7

Page 18: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Green500 on Nov. 2016

2017/04/05

1st F-J-G Workshop, Tokyo18

# # HPL Machine Architecture Country Rmax (TFLOPS) MFLOPS/W

1 28 DGX SaturnV GPU cluster (NVIDIA DGX1) USA 3,307.0 9462.1

2 8 Piz Daint, CSCS MPP (Cray, XC50:CPU + GPU) Switzerland 9,779.0 7453.5

3 116 Shoubu PEZY ZettaScaler-1 Japan 1,001.0 6673.8

4 1 TaihuLight MPP (Sunway SW26010) China 93,014.6 6051.3

5 375 QPACE3 Cluster (Fujitsu, KNL) Germany 447.1 5806.3

6 6 Oakforest-PACS, JCAHPC Cluster (Fujitsu, KNL) Japan 13,554.6 4985.1

7 18 Theta MPP (Cray XC40, KNL) USA 5,095.8 4688.0

8 162 XStream MPP (Cray CS-Storm, GPU) USA 781.3 4112.1

9 33 Camphor2 MPP (Cray XC40, KNL) Japan 3,057.4 4086.8

10 397 SciPhi XVI Cluster (KNL) USA 425.9 3836.7

Page 19: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

HPCG on Nov. 2016

2017/04/05 1st F-J-G Workshop, Tokyo 19

Page 20: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

McKernel support

n McKernel (special light weight kernel for Many-Core architecture)n developed at U. Tokyo and now at AICS, RIKEN (lead by Y.

Ishikawa)n KNL-ready version is almost completedn It can be loaded as a kernel module to Linuxn Batch scheduler is noticed to use McKernel by user’s script,

then applies itn Detach the McKernel module after job execution

2017/04/05

1st F-J-G Workshop, Tokyo20

Page 21: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

XcalableMP (XMP) support

n XcalableMP: massively parallel description language based on PGAS model and directives by usersn originally developed at U. Tsukuba and now at AICS, RIKEN

(lead by M. Sato)n KNL-ready version is under evaluation and tuningn It will be open for users to write (relatively) easy-way for large

scale parallelization as well as performance tuning

2017/04/05

1st F-J-G Workshop, Tokyo21

Page 22: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Memory Model (currently planned)n Our challenge – semi-dynamic switching of CACHE and FLAT

modesn Initial: nodes in the system are configured with a certain ratio of mixture

(half and half) of Cache and Flat modesn Batch scheduler is noticed about the memory configuration from user’s

scriptn Batch scheduler tries to find appropriate nodes without reconfigurationn If there are not enough number of nodes, some of them are rebooted with

another memory configurationn Reboot is by warm-reboot, with ~100 nodes groupn Size limitation (max. # of nodes) may be applied

n NUMA modeln Currently quadrant mode onlyn (Perhaps) we will not dynamically change it ??

2017/04/05

1st F-J-G Workshop, Tokyo22

Page 23: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

System operation outline

n Regular operationn both universities share the CPU time based on the budget ration not split the system hardware, but split the “CPU time” for flexible

operation (except several specially dedicated partitions)n single system entry for HPCI program, and other own program by each

university is performed under “CPU time” sharing

n Special operationn (limited period) massively large scale operation

-> effectively using the largest class resource in Japan for special occasion (ex. Gordon Bell Challenge)

n Power saving operationn power capping feature for energy saving scheduling feature reacts to

power saving requirement (ex. summer time)

2017/04/051st F-J-G Workshop, Tokyo23

Page 24: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

OFP resource sharing program (nation-wide)

n JCAHPC (20%)n HPCI – HPC Infrastructure program in Japan to share all supercomputers

(free!)n Big challenge special use (full system size)

n U. Tsukuba (23.5%)n Interdisciplinary Academic Program (free!)n Large scale general use

n U. Tokyo (56.5%)n General usen Industrial trial usen Educational usen Young & Female special use

2017/04/05

1st F-J-G Workshop, Tokyo24

Page 25: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Machine location: Kashiwa Campus of U. Tokyo

2017/04/05

1st F-J-G Workshop, Tokyo25

Hongo Campus of U. Tokyo

U. Tsukuba

KashiwaCampusof U. Tokyo

Page 26: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Xeon Phi tuning on ARTED (with Y. Hirokawa under collaboration with Prof. K. Yabana, CCS)

n ARTED – Ab-initio Real-Time Electron Dynamics simulatorn Multi-scale simulator based on RTRSDFT (Real-Time Real-Space Density

Functional Theory) developed in CCS, U. Tsukuba to be used for Electron Dynamics Simulationn RSDFT : basic status of electron (no movement of electron)n RTRSDFT : electron state under external force

n In RTRSDFT, RSDFT is used for ground staten RSDFT : large scale simulation with 1000~10000 atoms (ex. K-Computer)n RTRSDFT : calculate a number of unit-cells with 10 ~ 100 atoms

26

Journal of Advanced Simulation in Science and Engineering

Solids

Z

Electric field

y

Atom

x

Macroscopic grids Microscopic grids Vacuum

Figure 1: A schematic picture of the multi-scale coordinates system. Left-hand side showsthe macroscopic coordinate to describe propagation of the macroscopic electromagneticfields. Righ-hand side shows the microscopic coordinates to describe quantum dynamicsof electrons induced by the fields.

2. Theoretical formulation and scientific aspects

2.1. Theoretical framework

To explain our multi-scale simulation, we consider a simple case: a linearly polarized laserpulse irradiating normally on a surface of bulk Si. We take a coordinate system shown inFig. 1 where the surface of bulk Si is taken to be a Z = 0 plane, the direction of the laserelectric field is parallel to x-axis, and the direction of the laser propagation is parallel toz-axis. We denote the macroscopic coordinate in z-direction as Z. We describe the macro-scopic electromagnetic field of laser pulse using a vector potential, A⃗Z(t), which is relatedto the electric field by E⃗Z(t) = −(1/c)(dA⃗Z(t)/dt). The vector potential A⃗Z(t) satisfies thefollowing Maxwell equation:

1c2∂2

∂t2 A⃗Z(t) − ∂2

∂Z2 A⃗Z(t) =4πc

J⃗Z(t), (1)

where J⃗Z(t) is the electric current density at Z. The current is obtained from microscopiccalculation of electron dynamics as described below.

We next turn to the microscopic calculation of electron dynamics at each macroscopicpoint, Z. We use the TDKS equation for it. A symbol r⃗ is used to denote the microscopiccoordinates of electrons at macroscopic position Z. Since the wavelength of the laser pulseis much longer than the spatial scale of the electron dynamics in solids, we assume that themacroscopic electric field can be regarded as a spatially uniform field in the microscopicscale. We thus solve the following TDKS equation for electrons:

i∂

∂tubk⃗,Z (⃗r, t) =

!1

2m

"p⃗ + !k⃗ +

ec

A⃗Z(t)#2+ V$

ubk⃗,Z (⃗r, t), (2)

where ubk⃗,Z (⃗r, t) is the time-dependent Bloch orbital that has the same periodicity as that ofthe crystalline solid, ubk⃗,Z (⃗r, t) = ubk⃗,Z (⃗r+ a⃗, t). We employ a standard Kohn-Sham Hamilto-nian in the adiabatic local-density approximation in Eq. (2). A periodic potential V includes

100

2017/04/051st F-J-G Workshop, Tokyo

RSDFT : Real-Space Density Functional TheoryRTRSDFT : Real-Time RSDFT

Page 27: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Computation domain and amount

n Parameters for wave function expressionn k-points (NK), band-number (NB), 3-D lattice points (NL)n valuables are in double precision complex with matrix of (NK, NB, NL)n for stencil computation, size NL of calculation is performed NKxNB times

n Parameters used in this research (two models)n SiO2 : (43, 48, 36000 = (20, 36, 50)) -> not enough largen Si : (243, 32, 4096 = (16, 16, 16)) -> larger parallelism on thread

n NK is parallelized by MPI, then NKxNB is parallelized in OpenMPn domain of each process: (NK/NP, NB, NL)

(NP = number of processes)n space domain is not decomposed to minimize MPI communication

272017/04/05

1st F-J-G Workshop, Tokyo

Page 28: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Stencil computation (3D)

28

93.0157.6

251.4

547.0467.9

758.4

591.4690.3

0100200300400500600700800

KNC x2 KNL

Perf

orm

ance

[GFL

OPS

]

Si case

Original Compiler vec.

Explicit vec. (w/o SWP) Explicit vec. (w SWP)

57.2148.3185.0

442.0

230.6

542.9

336.4

593.8

0100200300400500600700800

KNC x2 KNL

Perf

orm

ance

[GFL

OPS

]

SiO2 case

Original Compiler vec.

Explicit vec. (w/o SWP) Explicit vec. (w SWP)

3x faster than KNC (at maximum)

2017/04/051st F-J-G Workshop, Tokyo

Page 29: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

KNL vs GPU

1st F-J-G Workshop, Tokyo

Peak performance (DP)

Actual memory bandwidth Actual B/F

Xeon Phi 7110P (KNC) 1074 GFLOPS 177.1 GB/s 0.16

Xeon Phi 7250 (KNL) 2998 GFLOPS 456.2 GB/s 0.15

Tesla K40 (Kepler) 1430 GFLOPS 180.5 GB/s 0.13

Tesla P100 (Pascal) 5300 GFLOPS 514.8 GB/s 0.10

29

Si case GFLOPS vs. Peak perf.

Xeon E5-2670v2 x2 (IVB) 232.1 58.0%

Xeon Phi 7110P x2 (KNC) 592.3 27.6%

OFP: Xeon Phi 7250 (KNL) 758.0 24.8%

Tesla K40 x2 (Kepler) 476.0 33.3%

Tesla P100 (Pascal) 788.2 14.9%

SiO2 case GFLOPS vs. Peak perf.

Xeon E5-2670v2 x2 (IVB) 239.2 59.8%

Xeon Phi 7110P x2 (KNC) 335.7 15.6%

OFP: Xeon Phi 7210 (KNL) 593.0 19.4%

Tesla K40 x2 (Kepler) 561.4 39.3%

Tesla P100 (Pascal) 856.6 16.2%

GPU (Pascal) performance is by courtesy of A. Naruse @ NVIDIA

2017/04/05

Page 30: Oakforest-PACS (OFP)events.science-japon.org/hpc17/slides/Taisuke Boku - FJG... · 2017-08-08 · 2008 2010 2012 2014 2016 2018 2020 U. of Tsukuba U. of Tokyo Kyoto U. Riken AICS

Center for Computational Sciences, Univ. of Tsukuba

Summaryn JCAHPC is a joint resource center for advanced HPC by U. Tokyo and U.

Tsukuba as the first case in Japan

n Oakforest-PACS (OFP) with 25 PFLOPS peak is ranked as #1 in Japan and #6 in the world, with Intel Xeon Phi (KNL) and OPA

n Under JCAHPC, both universities perform nation-wide resource sharing programs including HPCI

n JCAHPC is not just an organization to manage the resource but also a basic community for advanced HPC research

n OFP is used not only for HPCI and other resource sharing program but also a testbed for McKernel and XcalableMP system software to support Post-K development

2017/04/051st F-J-G Workshop, Tokyo30


Recommended