+ All Categories
Home > Documents > RAL and UK Plans and Outlook

RAL and UK Plans and Outlook

Date post: 03-Jan-2016
Category:
Upload: lionel-duffy
View: 45 times
Download: 0 times
Share this document with a friend
Description:
RAL and UK Plans and Outlook. LCG Service Challenge Meeting January 27-28, 2005 Rutherford Appleton Laboratory. Multiple Experiments. ATLAS. LHCb. CMS. SAMGrid (FermiLab). BaBar (SLAC). QCDGrid. PhenoGrid. UK Tier-1/A Centre Rutherford Appleton Laboratory. Grid Resource Discovery - PowerPoint PPT Presentation
16
John Gordon Deputy Director, CCLRC e-Science Centre [email protected] RAL and UK Plans and Outlook LCG Service Challenge Meeting January 27-28, 2005 Rutherford Appleton Laboratory
Transcript
Page 1: RAL and UK Plans and Outlook

John Gordon

Deputy Director,

CCLRC e-Science Centre

[email protected]

RAL and UK Plans and Outlook

LCG Service Challenge Meeting

January 27-28, 2005

Rutherford Appleton Laboratory

Page 2: RAL and UK Plans and Outlook

John Gordon

E-Science centre

Page 3: RAL and UK Plans and Outlook

John Gordon

E-Science centre

Multiple Experiments

ATLAS LHCb CMS

BaBar (SLAC) SAMGrid (FermiLab)QCDGrid PhenoGrid

Page 4: RAL and UK Plans and Outlook

John Gordon

E-Science centre

UK Tier-1/A Centre Rutherford Appleton Laboratory

• High quality data services

• National and international role

• UK focus for international Grid development

1000 CPU200 TB Disk60 TB Tape

(Capacity 1PB)

Grid Resource Discovery Time = 8 Hours

2004 CPU Utilisation2004 Disk Use

Page 5: RAL and UK Plans and Outlook

10 January 2005 Oversight Committee Tony Doyle - University of Glasgow

LHCb Data Challenge

424 CPU years (4,000 kSI2k months), 186M events • UK’s input significant (>1/4 total) • LCG(UK) resource:

– Tier-1 7.7%– Tier-2 sites:– London 3.9%– South 2.3%– North 1.4%

• DIRAC:– Imperial 2.0%– L'pool 3.1%– Oxford 0.1%– ScotGrid 5.1%

DIRAC alone

LCG inaction

1.8 106/day

LCG paused

Phase 1 Completed

3-5 106/dayLCG

restarted

186 M Produced Events

Entering Grid Production Phase..

Page 6: RAL and UK Plans and Outlook

10 January 2005 Oversight Committee Tony Doyle - University of Glasgow

Data Challenge

ATLAS DC2 - LCG - September 71%

2%

0%

1%

2%

14%

3%

1%

3%

9%

8%

3%2%5%1%4%

1%

1%

3%

0%

1%

1%

4%1%

0%

12%

0%

1%

1%

2%

10%

1% 4%

at.uibk

ca.triumf

ca.ualberta

ca.umontreal

ca.utoronto

ch.cern

cz.golias

cz.skurut

de.fzk

es.ifae

es.ific

es.uam

fr.in2p3

it.infn.cnaf

it.infn.lnl

it.infn.mi

it.infn.na

it.infn.na

it.infn.roma

it.infn.to

it.infn.lnf

jp.icepp

nl.nikhef

pl.zeus

ru.msu

tw.sinica

uk.bham

uk.ic

uk.lancs

uk.man

uk.rl

ATLAS DC2 - CPU usage

LCG41%

NorduGrid30%

Grid329%

LCG

NorduGrid

Grid3

Total:

~ 1350 kSI2k.months~ 95000 jobs~ 7.7 Million events fully simulated (Geant4)~ 22 TB

• 7.7 M GEANT4 events and 22 TB• UK ~20% of LCG• Ongoing..

• (3) Grid Production

• ~150 CPU years so far

• Largest total computing requirement

• Small fraction of what ATLAS need..

Entering Grid Production Phase..

Page 7: RAL and UK Plans and Outlook

John Gordon

E-Science centre

Capacity Planning

Table of Estimated Shares. Units are in TB and kSI2k as appropriate.                  

September 2 2005 2006 2007 2008 2009 2010  

 Dis

k CPUTap

eDis

kCPU TapeDis

kCPU Tape Disk CPU Tape Disk CPU Tape Disk CPU Tape  

Available 457 1508 882 748 2456 1500131

04276 2100 2087 6631 2100 3557 10163 2100 3205 9143 2100  

ALICE 6 19 20 11 43 23 22 94 35 35 142 61 59 218 61 53 196 61  

ATLAS 124 436 174 274 647 571 550 1439 805 987 2412 942 1682 3697 942 1516 3326 942  

CMS 95 100 168 111 140 418 212 279 657 442 868 491 754 1330 491 680 1197 491  

LHCb 20 213 121 38 476 138 74 1031 208 175 1734 79 299 2658 79 269 2391 79  

LCG Total 245 768 483 434 1150 1150 858 2843 1639 1639 5156 1573 2794 7903 1573 2518 7110 1573  

BaBar 84 348 149 93 389 59 101 423 30 102 380 24 174 582 24 157 523 24  

CDF 17 83 0 32 160 0 79 160 0 74 145 0 127 222 0 114 200 0  

D0 45 35 49 78 60 55 121 34 63 113 31 66 193 47 66 174 43 66  

H1 42 44 116 69 87 118 84 117 111 78 106 115 134 163 115 121 147 115  

MINOS 3 43 72 4 81 104 7 152 179 11 230 311 18 352 311 16 317 311  

SNO 19 25 13 32 50 14 35 61 11 33 55 12 56 85 12 50 76 12  

ZEUS 0 126 0 0 250 0 0 339 0 0 307 0 0 470 0 0 423 0  

PhenoGrid 0 11 0 1 26 0 1 56 0 2 84 0 3 128 0 2 116 0  

UKQCD 2 23 0 7 48 0 22 91 0 35 137 0 59 210 0 53 189 0  

Total 457 1508 882 748 2456 1500131

04276 2100 2087 6631 2100 3557 10163 2100 3205 9143 2100  

Still in the Planning Phase.

Not to be used elsewhere

Still in the Planning Phase.

Not to be used elsewhere

Page 8: RAL and UK Plans and Outlook

John Gordon

E-Science centre

Tier1 Planning

• Planning tool to balance requirements, finance, and technology foresight

• This gives plan for how we share the resources• Turned into 1/4ly allocations by UserBoard• Used in allocating disk and batch shares

– cpu isn’t a hard limit, just guidance to the batch scheduler

– Tier1A Manager and UB chair can vary in response to circumstances

Page 9: RAL and UK Plans and Outlook

John Gordon

E-Science centre

UK Tier-2 Centres

ScotGridDurham, Edinburgh, Glasgow NorthGridDaresbury, Lancaster, Liverpool,Manchester, Sheffield

SouthGridBirmingham, Bristol, Cambridge,Oxford, RAL PPD, Warwick

LondonGridBrunel, Imperial, QMUL, RHUL, UCL

Page 10: RAL and UK Plans and Outlook

Presenter Name

Facility Name

Tier-2 Board

Tier-2 Board Membership

Chair: Steve Lloyd4 Tier-2 Regional Management Board Chairs:Dave Colling, Roger Jones, Phil Clark, Jeff Tseng4 Tier-2 Technical Coordinators:Owen Maroney, Alessandra Forti, Fraser Speirs, Rhys NewmanProduction Manager: Jeremy ColesTier-1 Board Representative: John GordonUser Board Representative: tbdDeployment Board Representative: Dave Kelsey

Page 11: RAL and UK Plans and Outlook

10 January 2005 Oversight Committee Tony Doyle - University of Glasgow

Experimental Shares

Institute FacilityNo of CPUs

Total CPU

(KSI2000)

Total Formatte

d Disk (TB)

HEP Shar

e

Nominal Experiment Shares

ALICE ATLAS CMS LHCb BaBar CDF D0 Other

QMUL E-Science Cluster 344 354.0 36.0 75%   50%     15%     35%

QMUL BaBar Farm 80 33.0 1.0 100%         100%      

Attempting to determine what the ‘fair shares’ of the Tier-2 resources are.

Linear Collider

Multiply each resource by HEP Share and experiment share (+ time/belief weight) and integrate over all institutes.

Page 12: RAL and UK Plans and Outlook

10 January 2005 Oversight Committee Tony Doyle - University of Glasgow

Experimental Shares

CPU Disk

ALICE 2% 0%

ATLAS 37% 42%

CMS 7% 4%

LHCb 11% 13%

BaBar 12% 14%

CDF 3% 4%

D0 12% 14%

Other 16% 9%

Preliminary!

Raw data needs to be checked/updated

Page 13: RAL and UK Plans and Outlook

10 January 2005 Oversight Committee Tony Doyle - University of Glasgow

Available Hardware Summary

• Survey of Tier2 provided commitments of resources to GridPP

CPU (KSI2K

)Disk (TB)

London 1820 139

NorthGrid 2602 543

ScotGrid 271 102

SouthGrid 701 49

Page 14: RAL and UK Plans and Outlook

Presenter Name

Facility Name

T2 Experiment SharesCPU (KSI2K) Disk (TB)

2004 2005 2006 2007 2004 2005 2006 2007

ALICE 75 85 108 109 12 13 15 15

ATLAS 1661 1874 2389 2412 275 295 335 338

CMS 313 353 450 454 52 56 63 64

LHCb 468 528 673 680 77 83 94 95

BaBar 507 571 728 735 84 90 102 103

CDF 162 183 233 236 27 29 33 33

D0 507 572 730 737 84 90 102 103

Other 703 793 1011 1021 116 125 142 143

Total 4397 4960 6322 6382 728 781 887 894

• We have defined shares but applying them across T2s will be more complex

Page 15: RAL and UK Plans and Outlook

John Gordon

E-Science centre

Issues

• Moores Law and other hardware unknowns• Long lead times on Mass Storage acquisitions• Overcomplication, reliability and maintainability of

Middleware• Conflict between experiment demands

– and understanding the computing models • Balance between 2008 and 2005-7

– buying hardware that isn’t used weakens the case for the hardware that will be required in 2008, also true for networking

• T1-T2 Issues – still needs investigation

Page 16: RAL and UK Plans and Outlook

John Gordon

E-Science centre

Summary

• UK has a working Tier1

– Engaged with (most) experiments and contributing to DCs

• UK has an impressive range of Tier2 resources available

– A good fraction are on the grid• Challenges

– Improve quality of integrated management– Copying with middleware– Keep up with the increasing demands


Recommended