Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

15
Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008

Transcript of Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

Page 1: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

Grid Computing

Oxana Smirnova

NDGF- Lund University

R-ECFA meeting in Sweden

Uppsala, May 9, 2008

Page 2: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

2

Computing challenges at LHC

2008-05-09

Page 3: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

3

“Full chain” of HEP data processing

2008-05-09

Event generation (Pythia)

Detector simulation (Geant4)

Hit digitization100011110101110101100101110110100

Reconstruction

Analysis data preparation

Analysis, results (ROOT)

Slide adapted from Ch.Collins-Tooth and J.R.Catmore

Page 4: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

4

ATLAS Monte Carlo data production flow (10 Mevents)

2008-05-09

HitsMCTruth

Digits(RDO)

MCTruth

BytestreamRaw

Digits

ESD

ESD

Geant4

Reconstruction

Reconstruction

Pile-up

BytestreamRaw

Digits

BytestreamRaw

Digits

HitsMCTruth

Digits(RDO)

MCTruth

Physicsevents

EventsHepMC

EventsHepMC

HitsMCTruth

Digits(RDO)

MCTruthGeant4

Geant4

Digitization

Digits(RDO)

MCTruth

BytestreamRaw

Digits

BytestreamRaw

Digits

BytestreamRaw

DigitsEventsHepMC

HitsMCTruth

Geant4Pile-up

Digitization

Mixing

Mixing Reconstruction ESD

Pyt

hia

Event generation

DetectorSimulation

Digitization(Pile-up)

ReconstructionEventMixingByte stream

EventsHepMC

Min. biasEvents

Piled-upevents Mixed events

Mixed eventsWith

Pile-up

~2 TB 24 TB 75 TB18 TB 5 TB

TBVolume of datafor 107 events

Persistency:Athena-POOL

• Very different tasks/algorithms (ATLAS experiment in this example)• Single “job” lasts from 10 minutes to 1 day• Most tasks require large amounts of input and produce large output data

• Very different tasks/algorithms (ATLAS experiment in this example)• Single “job” lasts from 10 minutes to 1 day• Most tasks require large amounts of input and produce large output data

Page 5: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

5

LHC computing specifics

Data-intensive tasks Large datasets, large files Lengthy processing times Large memory consumption High throughput is necessary

Very distributed computing and storage resources CERN can host only a small

fraction of needed resources and services

Distributed computing resources of modest size

Produced and processed data are hence distributed, too

Issues of coordination, synchronization, data integrity and authorization are outstanding

2008-05-09

Page 6: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

6

Software for HEP experiments

Massive pieces of software• Written by very many different authors in different languages (C++, Java, Python, Fortran)• Dozens of external components• Occupy as much as ~10 GB of disk space each release

Frequent releases

• Every experiment produces a release as often as once a month during the preparation phase (which is now for LHC)

Difficult to set up outside the lab

• Experiments can not afford supporting different operating systems and different computer configurations

For a small university group it is very difficult to manage different software sets and maintain hardware

•ALICE, ATLAS, PHENIX etc – all in many versions

Solution: use the Grid

2008-05-09

Page 7: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

7

Grid is a result of IT progress

2008-05-09

Graph from “The Triumph of the Light”, G. Stix, Sci. Am. January 2001

•Computer speed doubles every 18 months•Network speed doubles every 9 months

Network vs.

computer performan

ce:

•Computers: 500 times faster•Networks: 340000 times faster

1986 to

2000:

•Computers: 60 times faster•Networks: 4000 times faster

2001 to 2010

(projected):

Page 9: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

9

Grids in LHC experiments

2008-05-09

Almost all Monte Carlo and data processing today is done via Grid

There are 20+ Grid flavors out there Almost all are tailored for a specific application and/or specific

hardware LHC experiments make use of 3 Grid middleware flavors:

gLite ARC OSG

All experiments develop own higher-level Grid middleware layers ALICE – AliEn ATLAS – PANDA and DDM LHCb – DIRAC CMS – ProdAgent and PhEDEx

Page 10: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

10

ATLAS Experiment at CERN - Multi-Grid Infrastructure

2008-05-09

Graphics from a slide by A.Vaniachine

Page 11: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

Nordic DataGrid Facility

2008-05-09 11

Provides a unique distributed “Tier1” center via NorduGrid/ARC

Involves 7 largest Nordic academic HPC centers

…plus a handful of University centers (Tier2 service)

Connected to CERN directly with GEANT 10GBit fiber

Inter-Nordic shared 10Gbit network from NORDUnet

Budget: staff only, 2 MEUR/year, by Nordic research councils

Page 12: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

12

Swedish contribution: SweGrid

Investment Time Cost, KSEK

Six clusters (6x100 cores) including 12 TB FC disk Dec 2003 10 173

Disk storage part 1, 60 TB SATA May 2004 2 930

Disk storage part 2, 86.4 TB SATA May 2005 2 119

2008-05-09

Centre Tape volume, TB

Cost, KSEK

HPC2N 120 1000

PDC 120 1000

NSC 120 1000

SweGrid in 2003-2007Location Profile

HPC2N (Umeå) IT

UPPMAX (Uppsala) IT, HEP

PDC (Stockholm) IT

C3SE (Gothenburg) IT

NSC (Linköping) IT

Lunarc (Lund) IT, HEP

Co-funded by the Swedish Research Council and the Knut and Alice Wallenberg foundation

One technician per center Middleware: ARC, gLite 1/3 allocated to LHC Computing

Page 13: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

13

SweGrid and NDGF usage

2008-05-09

LHC ChemistryOther Physics Amanda/IcecubeGeo Science Fluid MechanicsBioinformatics MedicineComputer Science BiotechnologyMathematics StatisticsPharmacology Material ChemistryElectronics

SWEGRID USAGE

ATLAS PRODUCTIONIN 2007

Page 14: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

14

Swedish contribution to LHC-related Grid R&D

NorduGrid (Lund, Uppsala, Umeå, Linköping, Stockholm and others) Produces ARC middleware, 3 core developers are in Sweden

SweGrid: tools for Grid accounting, scheduling, distributed databases Used by NDGF, other projects

NDGF: interoperability solutions EU KnowARC (Lund, Uppsala + 7 partners)

3 MEUR project (3 years), develops next generation ARC. Project’s technical coordinator is in Lund

EU EGEE (Umeå, Linköping, Stockholm)

2008-05-09

Page 15: Grid Computing Oxana Smirnova NDGF- Lund University R-ECFA meeting in Sweden Uppsala, May 9, 2008.

15

Summary and outlook

Grid technology is vital for the success of LHC Sweden contributes very substantially with hardware,

operational support and R&D Very high efficiency

Sweden has signed MoU with LHC Computing Grid in March 2008 Pledge of long-term computing service for LHC

SweGrid2 is coming A major upgrade of SweGrid resources Research Council granted 22.4 MSEK for investments and

operation in 2007-2008 43 MSEK more are being requested for years 2009-2011 Includes not just Tier1, but also Tier2 and Tier3 support

2008-05-09