NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader [email protected]...

21
NERSC Workload Analysis John Shalf SDSA Team Leader [email protected] NERSC User Group Meeting September 17, 2007

Transcript of NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader [email protected]...

Page 1: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC Workload Analysis

John ShalfSDSA Team Leader

[email protected]

NERSC User Group MeetingSeptember 17, 2007

Page 2: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 1

Workload Analysis

• Purpose– Understand NERSC User Requirements– Prioritize software library and tool

optimization and deployment– Inform NERSC Sustained System

Performance (SSP) Benchmark Selection• “Benchmarks are only useful insofar as they

model the intended computational workload.”Ingrid Bucher & Joanne Martin, LANL, 1982

• Effective performance on SSP to reflect effectiveperformance on NERSC workload

Page 3: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 2

Balancing Requirements

• NERSC Workload overview– ~3000 users– ~300 projects respresenting a broad range of science– ~700 codes (>2 codes per project on average!)– 15 science areas for 6 Office of Science divisions

• Select a subset (<10) codes to represent therequirements of the workload– Weight based on code’s contribution workload?

(workload coverage)– Weight equally for each area of science?

(algorithm/science-area coverage)

• Attempt to cover both dimensions– Still daunting– Search for islands of coherence in the codes or

algorithm selection by different scientific disciplines

Page 4: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 3

Workload OverviewDOE Ofice

Science Area Data ASCR BER BES FES HEP NP Grand TotalAccelerator Physics Normalized Sum of MPP Request 1020000 9768000 758000 11546000

MPP Request Count 3 32 9 44Applied Math Normalized Sum of MPP Request 10107000 10000 10117000

MPP Request Count 23 1 24Astrophysics Normalized Sum of MPP Request 9888000 2602000 12490000

MPP Request Count 42 6 48Chemistry Normalized Sum of MPP Request 20000 22534000 22554000

MPP Request Count 1 106 107Climate Research Normalized Sum of MPP Request 400000 15066000 15466000

MPP Request Count 1 44 45Computer Sciences Normalized Sum of MPP Request 1400000 1400000

MPP Request Count 41 41Engineering Normalized Sum of MPP Request 20000 1910000 1930000

MPP Request Count 2 7 9Environmental Sciences Normalized Sum of MPP Request 170000 170000

MPP Request Count 3 3Fusion Energy Normalized Sum of MPP Request 350000 40200990 40550990

MPP Request Count 2 135 137Geosciences Normalized Sum of MPP Request 2704000 2704000

MPP Request Count 15 15High Energy Physics Normalized Sum of MPP Request 500000 500000

MPP Request Count 1 1Lattice Gauge Theory Normalized Sum of MPP Request 16910000 17463000 34373000

MPP Request Count 9 5 14Life Sciences Normalized Sum of MPP Request 50000 11019000 40000 11109000

MPP Request Count 1 27 1 29Materials Sciences Normalized Sum of MPP Request 570000 22640750 23210750

MPP Request Count 3 131 134Nuclear Physics Normalized Sum of MPP Request 50000 12696000 12746000

MPP Request Count 1 31 32Total Normalized Sum of MPP Request 12967000 26255000 50858750 40200990 37066000 33519000 200866740Total MPP Request Count 75 74 264 135 84 51 683

Page 5: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 4

ERCAP Allocations 2007By Office and Science Area

Awards by DOE Office

Advanced Scientific Computing Research

(ASCR)4%

Basic Energy Research (BER)14%

Basic Energy Science (BES)29%

Fusion (FES)29%

High Energy Physics (HEP)15%

Nuclear Physics (NP)9%

Awards By Science Area

Accelerator Physics7%

Applied Math2%

Astrophysics6%

Chemistry13%

Climate Research9%

Computer Sciences1%

Engineering1%

Environmental Sciences0%

Fusion Energy28%

Geosciences2%

High Energy Physics0%

Latice Gauge Theory7%

Life Sciences6%

Materials Sciences13%

Nuclear Physics5%

Page 6: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 5

Focus on Science Areas

Astrophysics6%

Accelerator Physics7%

Lattice Gauge Theory7%

Climate Research9%

Materials Science13%

Chemistry13%

Fusion Energy28%

Life Sciences6%

Environmental Sciences0%

Engineering1%

Geosciences2%

Computer Science1%

Applied Math2%

Nuclear Physics5%

High Energy Physics0%

Page 7: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 6

Focus on Science Areas

Astrophysics6%

Accelerator Physics7%

Lattice Gauge Theory7%

Climate Research9%

Materials Science13%

Chemistry13%

Fusion Energy28%

Life Sciences6%

Environmental Sciences0%

Engineering1%

Geosciences2%

Computer Science1%

Applied Math2%

Nuclear Physics5%

High Energy Physics0%

Page 8: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 7

Modeling the NERSC Workload

• Target: choose 6 representative applicationsfrom 288 projects, 684 code descriptions

Science Category Award MPP Hours Number of Projects

Fusion Energy 16,509,000 51Chemistry 7,405,000 41Materials Science 7,385,000 58Climate Research 4,911,000 29Lattice Gauge Theory 3,812,500 6Accelerator Physics 3,650,000 16Astrophysics 3,319,000 16Life Sciences 3,149,000 20Nuclear Physics 2,790,000 16Applied Math 930,000 10Geosciences 920,000 6Computer Science 731,500 13Engineering 390,000 4High Energy Physics 70,000 1Environmental Sciences 24,000 1

Page 9: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 8

Climate Modeling (BER)

Climate with INCITE

GCM1%

ATHAM1% POP

1%WRF1%

IMPACT3%

GCRM4%

CCSM28%

Filter24%

Forecast24%

CAM13%

Climate without INCITE

CAM24%

CCSM55%

WRF1%

POP1%

ATHAM2%

GCM2%

IMPACT6%

GCRM9%

Page 10: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 9

Climate Modeling (BER)

Climate without INCITE

CAM24%

CCSM55%

WRF1%

POP1%

ATHAM2%

GCM2%

IMPACT6%

GCRM9%

Code MPP Award Percent Cumulative%

1 CCSM 2,342,000 51% 51%2 CAM 2,000,000 23% 74%3 GCRM 2,000,000 8% 82%4 IMPACT 1,085,000 6% 88%5 GCM 375,000 2% 90%6 ATHAM 280,000 2% 92%7 POP 100,000 1% 93%8 WRF 80,000 1% 94%

•CAM and POP dominate CCSMcomputational requirements

•FV-CAM increasingly replacing Spectral-CAM in future CCSM calculations

•FV-CAM with D-Mesh selected(coordinate w/NCAR procurement)

Page 11: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 10

Material Science

QBox3%SIESTA

5%RGWBS

3%

PEscan3%

PARATEC4%

PARSEC2%

PWscf2%

Glotzilla2%

cmat_atomistic2%

GW1%

ALCMD2%

BO-LSD-MD2%

TranG991%

SSEqmc1%

DFT1%

AFQMC1%

OLCAO1%

Moldy1%

Chebyshev1%

TRANSPORT0%

NWChem1%

BSE1%

BSE1%

becmw1%

NAMD1%

PEtot1%

CHAMP1%

NEMO 3D1%

CP1%

Planewave codes1%

SCARLET1%

QMhubbard1%

CF Monte-Carlo1%

sX-PEtot1%LS3DF

1%TBMD1%

DL_POLY0%

XqmmmX0%

LAMMPS0%

Real space multigrid0%

flair0%

WIEN2K0%GCMC0%

mol_dyn0%MC0%

FDTDGA0%

mxmat0%mxci0%

freepar0%

CL/GCMD0%

ESPRESSO0%

Tmatrix0%

Smatrix0%

MomMeth0%

FDTD5130%BEST0%

HOLLICITA0%

FEFF_OPCONS0%AndyS

0%ABINIT-DW

0%ARPES

0%NBSE-ABINIT

0%FEFFMPI

0%Hartree

0%

CASINO 4%

FLAPW, DMol3 5%

VASP26%

LSMS8%

GINGER0%

• 7,385,000 MPP hoursawarded

• 62 codes, 65 users

• Typical code used in 2.15allocation requests

Code MPP Hours Percent Cumulative%

1 VASP 1,992,110 26% 26%2 LSMS 600,000 8% 34%

3FLAPW, DMol3 350,000 5% 39%

4 CASINO 312,500 4% 43%5 QBox 262,500 3% 46%6 SIESTA 346,500 5% 51%7 RGWBS 232,500 3% 54%8 PEscan 220,000 3% 57%9 PARATEC 337,500 4% 61%

10 PARSEC 182,500 2% 64%Other 167,300 34% 66%

Page 12: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 11

Material Science

• 7,385,000 MPP hoursawarded

• 62 codes, 65 users

• Typical code used in 2.15allocation requests

Code MPP Hours Percent Cumulative%

1 VASP 1,992,110 26% 26%2 LSMS 600,000 8% 34%

3FLAPW, DMol3 350,000 5% 39%

4 CASINO 312,500 4% 43%5 QBox 262,500 3% 46%6 SIESTA 346,500 5% 51%7 RGWBS 232,500 3% 54%8 PEscan 220,000 3% 57%9 PARATEC 337,500 4% 61%

10 PARSEC 182,500 2% 64%Other 167,300 34% 66%

RGWBS3%

PEscan3%

PARATEC4%

PARSEC2%

FLAPW, DMol35%

CASINO4%

QBox3%

SIESTA5%

LSMS8%

Other (52 codes)37%

VASP26%

Page 13: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 12

Materials Science(by algorithm)

Multiple Scattering DFT17%

Localized Basis DFT12%

GW+BSE7%

Quantum Monte Carlo7%

Maffin Tin Sphere DFT5%

Classical force field Monte Carlo3%

Older PDE3%

Classical MD6%

Plane Wave DFT38%

Real Space DFT2% Analysis by Lin-Wang Wang

Page 14: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 13

Materials Science(by algorithm category)

GW+BSE7%

Quantum MC7%

Classical MD6%

Classical MC3%

Other PDE5%

Density Functional Theory (DFT)

72%

Analysis by Lin-Wang Wang

Page 15: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 14

Materials Science(by algorithm category)

GW+BSE7%

Quantum MC7%

Classical MD6%

Classical MC3%

Other PDE5%

Density Functional Theory (DFT)

72%

• Density Functional Theorycodes– >70% of the workload!

– Majority are planewave DFT!

• Common requirements for DFT– 3D global FFT

– Dense Linear Algebra fororthogonalization of wave basisfunctions

– Dense Linear Algebracalculating pseudopotential

• Dominant Code: VASP

• Similar Codes (planewave DFT)– QBox

– PARATEC

– PETOT/PESCAN

Page 16: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 15

Astrophysics

MADCAP24%

ENZO9%

SEDONA9%

VULCAN7%

SNIA7%

SNE7%

PHOENIX6%

V2D3%

CLOUDY3%

XING3%

TILLAMOOK3%

TREEPM2%

PM2%

NEK50002%

NVWA2%

DYNAMO2%

RAM2%

SRHD32%

AMR2%

TRISTAN2%

ART1%

SNOOP1%GADGET1%SUNRISE0%

CACTUS0%

• MADCAP: CMB Analysis Suite– Dominates allocations even

though it is not INCITE

– I/O dominated: Now covered inseparate I/O benchmarkingtests

• ENZO: INCITE AMR code

• SciDAC Astrophysics Codesdominant– Coverage of MHD + combustion

– Suggested to look at codes thatuse implicit methods ratherthan explicit timestepping(better representative of futurecodes)

– Might help with Fusioncoverage

Page 17: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 16

Other Application Areas

• Fusion: 76 codes– 5 codes account for >50% of workload:

OSIRIS, GEM, NIMROD, M3D, GTC– Further subdivide to PIC (OSIRIS, GEM, GTC)

and MHD (NIMROD, M3D) code categories

• Chemistry: 56 codes for 48 allocations– INCITE award (S3D) eclipses other chemistry

codes: put in separate category• Planewave DFT: VASP, CPMD, DACAPO• Quantum Monte Carlo: ZORI• Ab-initio Quantum Chemistry: Molpro, Gaussian,

GAMESS

– Planewave DFT dominates (but alreadycovered in MatSci workload)

– Small allocations Q-Chem category add up todominant workload component

• Accelerator Modeling– 50% of workload consumed by 3 codes

VORPAL, OSIRIS, QuickPIC– Dominated by PIC codes

code MPP Award Percent Cumulative%OSIRIS 2,112,500 11% 11%GEM 2,058,333 11% 22%NIMROD 2,229,167 12% 34%M3D 1,921,667 10% 45%GTC 1,783,333 10% 54%

Code Award Percent Cumulative%ZORI 695,000 12% 12%MOLPRO 519,024 9% 21%DACAPO 500,000 9% 29%GAUSSIAN 408,701 7% 36%CPMD 396,607 7% 43%VASP 371,667 6% 49%GAMESS 364,048 6% 56%

Code MPP Award Percent Cumulative%

VORPAL 1,529,786 33% 33%OSIRIS 784,286 16% 49%QuickPIC 610,000 13% 62%Omega3p 210,536 4% 66%Track3p 210,536 4% 70%

Page 18: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 17

Selecting Benchmarks

• Coverage– Cover science areas

– Cover algorithm space

• Portable– Robust ‘build’ systems

– Not architecture specific implementation

• Scalable– Do not want to emphasize applications that do not

justify scalable HPC resources

• Distributable– No proprietary or export-controlled code

• Availability of Developer for Assistance/Support

Page 19: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 18

Narrowing Selection

First Cut (Primary Secondary Kernel):Fusion PIC GEM (4.9) / GTC (4.4) / XGC (3.7) / SUMMIT (1.5)Acceleratormodeling

OSIRIS (2.9) / VORPAL (2.6) / IMPACTZT / BeamB3D /QuickPIC

Fusion MHD M3D / NIMROD / GS2 / GYRO / BOUT / DynamoDFT VASP / LSMS / PWscf / CPMD / PARATEC /Quantum Chem NWChem / SIESTA / CASINO / GAMESS / GaussianClimate CAM / Forecast Model / IMPACTLG Physics MILC / RHMCOther astro MADCAP / SN1A / VULCAN / SRH3D /

FLASH/MAESTROCombustion S3DLifeSci BLAST / Forge / RepeatMasker

Page 20: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 19

Narrowing Selection

First Cut (Primary Secondary Kernel):Fusion PIC GEM (4.9) / GTC (4.4) / XGC (3.7) / SUMMIT (1.5)Acceleratormodeling

OSIRIS (2.9) / VORPAL (2.6) / IMPACTZT / BeamB3D /QuickPIC

Fusion MHD M3D / NIMROD / GS2 / GYRO / BOUT / DynamoDFT VASP / LSMS / PWscf / CPMD / PARATEC /Quantum Chem NWChem / SIESTA / CASINO / GAMESS / GaussianClimate CAM / Forecast Model / IMPACTLG Physics MILC / RHMCOther astro MADCAP / SN1A / VULCAN / SRH3D /

FLASH/MAESTROLifeSci BLAST / Forge / RepeatMasker

Page 21: NERSC Workload Analysis€¦ · NERSC Workload Analysis John Shalf SDSA Team Leader jshalf@lbl.gov NERSC User Group Meeting September 17, 2007. NERSC User Group Meeting, September

NERSC User Group Meeting, September 17, 2007 20

Benchmark Summary

DDI, BLASF90# of electron/atoms

512Dense linearalgebra, DFT

Quantum Chem

(BES)

GAMESS

netCDFF90Nx, ny1024Navier StokesCFD, FFT

Climate (BER)CAM

LibrariesLangParametersProposedConcurrency

AlgorithmSpace

Science SpaceBenchmark

HDF5, Para-mesh

F90/C# of grid points1024, 2048PPM, AMR

Low MachHydro

Astrophysics(HEP)

FLASH ?

MAESTRO ?

H5Part(HDF5)

HDF5, aztec,lapack, petsc

F90/

C++

C/C++

# of particles,grid points

Largely PIC, fftcomponent

Finitedifference, PIC

AcceleratorPhysics (HEP)

IMPACTZ ?

VORPAL ?

ScalapackF90# of atoms512,

2048, 4096

DFTMat Science(BES)

PARATEC

NoneC# of grid points,steps pertrajectory

2048Conjugategradiant,sparse matrix

LG Physics (NP)MILC

FFT (opt)F90# of particlesand grid points

512, 2048PIC, finitedifference

Fusion (FES)GTC