ian bird lhc computing grid project leader lhc grid fest 3 rd october 2008 a worldwide collaboration

13
THE LHC GRID SERVICE Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

Upload: kathryn-day

Post on 05-Jan-2016

214 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

THE LHC GRID SERVICE

Ian Bird

LHC Computing Grid Project Leader

LHC Grid Fest3rd October 2008

A worldwide collaboration

Page 2: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

Introduction

• The LHC Grid Service is a worldwide collaboration between:– 4 LHC experiments and– ~140 computer centres that contribute resources– International grid projects providing software and services

• The collaboration is brought together by a MoU that:– Commits resources for the coming years– Agrees a certain level of service availability and reliability

• As of today 33 countries have signed the MoU:– CERN (Tier 0) + 11 large Tier 1 sites– 130 Tier 2 sites in 60 “federations”

• Other sites are expected to participate but without formal commitment

[email protected] 2

Page 3: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

The LHC Computing Challenge

[email protected] 3

Signal/Noise: 10-9

Data volume High rate * large number of

channels * 4 experiments 15 PetaBytes of new data each year

Compute power Event complexity * Nb. events *

thousands users 100 k of (today's) fastest CPUs 45 PB of disk storage

Worldwide analysis & funding Computing funding locally in major

regions & countries Efficient analysis everywhere GRID technology

Page 4: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

[email protected]

Tier 0 at CERN: Acquisition, First pass processing Storage & Distribution

4

1.25 GB/sec (ions)

Page 5: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

Tier 0 – Tier 1 – Tier 2

[email protected] 5

Tier-0 (CERN):•Data recording•Initial data reconstruction

•Data distribution

Tier-1 (11 centres):•Permanent storage•Re-processing•Analysis

Tier-2 (~130 centres):• Simulation• End-user analysis

Page 6: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

Evolution of Grids

[email protected] 6

Data Challenges

First physics

GRID 3

EGEE 1

LCG 1

EU DataGrid

GriPhyN, iVDGL, PPDG

EGEE 2

OSG

LCG 2

EGEE 3

1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008

WLCG Partially decentralized

model– replicate the event data at

about five regional centres

– data transfer via network ormovable media

RC2

CERN

RC1

CosmicsService Challenges

Page 7: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

CPU Usage Early 2008

CERNBNLTRIUMFFNALFZK-GRIDKACNAFCC-IN2P3RALASGCPICNDGFNL-T1Tier 2

Recent grid use• The grid concept really works – all

contributions – large & small contribute to the overall effort!

Tier 2: 54%

CERN: 11%

Tier 1: 35%

350k /day

Page 8: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

Data transfer out of Tier 0

• Full experiment rate needed is 650 MB/s

• Desire capability to sustain twice that to allow for Tier 1 sites to shutdown and recover

• Have demonstrated far in excess of that

• All experiments exceeded required rates for extended periods, & simultaneously

• All Tier 1s achieved (or exceeded) their target acceptance rates

Tier-2s and Tier-1s are inter-connected by the general

purpose research networks

Any Tier-2 mayaccess data at

any Tier-1

Tier-2 IN2P3

TRIUMF

ASCC

FNAL

BNL

Nordic

CNAF

SARAPIC

RAL

GridKa

Tier-2

Tier-2

Tier-2

Tier-2

Tier-2

Tier-2

Tier-2Tier-2

Tier-2

Page 9: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

Production Grids

• WLCG relies on a production quality infrastructure– Requires standards of:

• Availability/reliability• Performance• Manageability

– Will be used 365 days a year ... (has been for several years!)

– Tier 1s must store the data for at least the lifetime of the LHC - ~20 years• Not passive – requires active migration to newer media

• Vital that we build a fault-tolerant and reliable system– That can deal with individual sites being down and recover

[email protected] 10

Page 10: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

11

WLCG depends on two major science grid infrastructures ….

EGEE - Enabling Grids for E-ScienceOSG - US Open Science Grid ... as well as many national grid projects

Interoperability & interoperation is vital significant effort in building the procedures to support it

Page 11: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

Enabling Grids for E-sciencE

EGEE-II INFSO-RI-031688

240 sites45 countries45,000 CPUs12 PetaBytes> 5000 users> 100 VOs> 100,000 jobs/day

ArcheologyAstronomyAstrophysicsCivil ProtectionComp. ChemistryEarth SciencesFinanceFusionGeophysicsHigh Energy PhysicsLife SciencesMultimediaMaterial Sciences…

Apr-04Jul-0

4

Oct-04Jan

-05

Apr-05Jul-0

5

Oct-05Jan

-06

Apr-06Jul-0

6

Oct-06Jan

-07

Apr-07Jul-0

70

10000

20000

30000

40000

50000

No. CPU

Apr-04Jul-0

4

Oct-04Jan

-05

Apr-05Jul-0

5

Oct-05Jan

-06

Apr-06Jul-0

6

Oct-06Jan

-07

Apr-07Jul-0

70

50

100

150

200

250

300

No. Sites

Grid infrastructure project co-funded by the European Commission - now in 2nd phase with 91 partners in 32 countries

Page 12: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

NERSC

BU

UNM

SDSC

UTA

OU

FNAL

ANL

WISCBNL

VANDERBILT

PSU

UVA

CALTECH

IOWA STATE

PURDUE

IU

BUFFALO

TTU

CORNELL

ALBANY

UMICH

INDIANAIUPUI

STANFORD

UWM

UNL

UFL

UNI

WSUMSU

LTU

LSU

CLEMSONUMISS

UIUC

UCRUCLA

LEHIGH

NSF

ORNL

HARVARD

UIC

SMU

UCHICAGO

MIT

RENCI

LBL

GEORGETOWNUIOWA

UCDAVIS

ND

Partnering with: US LHC: Tier-1s, Tier-2s, Tier-3s Campus Grids: Clemson, Wisconsin, Fermilab, Purdue Regional & National Grids: TeraGrid, New York State Grid, EGEE, UK NGS International Collaboration: South America, Central America, Taiwan, Korea, UK.

Access to 45,000 Cores, 6 Petabytes Disk, 15 Petabytes Tape >15,000 CPU Days/Day

~85% Physics: LHC, Tevatron Run II, LIGO; ~15% non-physics: biology, climate, text mining, Including ~20% Opportunistic use of others resources.

Virtual Data Toolkit: Common software developed between Computer Science & applications used by OSG and others.

OSG Project :Supported by the Department of Energy & the National Science Foundation

Page 13: Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration

THE LHC GRID SERVICEA worldwide collaboration

Has been in production for several years

Is now being used for real data

Is ready to face the computing challenges as LHC gets up to full speed