+ All Categories
Home > Documents > Western Analysis Facility

Western Analysis Facility

Date post: 04-Jan-2016
Category:
Upload: julian-zamora
View: 20 times
Download: 0 times
Share this document with a friend
Description:
Western Analysis Facility. Richard P. Mount US-ATLAS Tier2/Tier3 Workshop University of Chicago Aug 20, 2009. Outline. SLAC HEP Computing Facilities SLAC Power and Cooling Infrastructure Hosting University Equipment SLAC Strengths Longer-Term Goals. SLAC HEP Computing Facilities. - PowerPoint PPT Presentation
21
Western Analysis Facility Richard P. Mount US-ATLAS Tier2/Tier3 Workshop University of Chicago Aug 20, 2009
Transcript
Page 1: Western Analysis Facility

Western Analysis Facility

Richard P. Mount

US-ATLAS Tier2/Tier3 Workshop

University of Chicago

Aug 20, 2009

Page 2: Western Analysis Facility

Outline

04/20/23 2Western Analysis Facility

• SLAC HEP Computing Facilities

• SLAC Power and Cooling Infrastructure

• Hosting University Equipment

• SLAC Strengths

• Longer-Term Goals

Page 3: Western Analysis Facility

04/20/23 3Western Analysis Facility

SLAC HEP Computing Facilities

54%

15%

9%

7%

2%

1%

0%6%

3%3%

SLAC HEP CPU 8600 Cores

BaBar

FGST

ATLAS

KIPAC

LCD

Klystron

Radiation Protection

Other

AllUsers

Beam Physics

72 Cores: SMP (SGI Altix)

360 Cores: Infiniband Cluster

128 Cores: Myrinet Cluster

8040 Cores: HEP Data Processing

SLAC HEP Batch CPU 8600 Cores

Plus 106 Interactive/ Build Cores

Page 4: Western Analysis Facility

04/20/23 4Western Analysis Facility

SLAC HEP Computing Facilities

0

100

200

300

400

500

600

700

800

900

2000 2001 2002 2003 2004 2005 2006 2007 2008 2009

CPU Box Age Distribution

0500

100015002000250030003500400045005000

2000 2001 2002 2003 2004 2005 2006 2007 2008 2009

CPU Core Age Distribution

Page 5: Western Analysis Facility

04/20/23 5Western Analysis Facility

SLAC HEP Computing Facilities

27%

35%

14%

12%

9%

3%

SLAC HEP Disk Space 2052 Terabytes

BaBar

FGST

ATLAS

KIPAC

Physics Groups

Other

34 TB: AFS

47 TB: NFS-Network Appliance

700 TB: NFS

1200 TB: xrootd

74 TB: Lustre, Objectivity …

Page 6: Western Analysis Facility

SLAC HEP Computing Facilities

0

10

20

30

40

50

60

2000 2001 2002 2003 2004 2006 2007 2008 2009

Storage Box Age Distribution

0

100

200

300

400

500

600

700

800

2000 2001 2002 2003 2004 2006 2007 2008 2009

Storage Terabyte Age Distribution

04/20/23 6Western Analysis Facility

Page 7: Western Analysis Facility

SLAC HEP Computing Facilities

Mass Storage Managed by HPSS

Upgrade in Progress

x 6 2 x

Transfer ~ 3 Petabytes

30,000 Slots Total 13,000 Slots Total

04/20/23 7Western Analysis Facility

Page 8: Western Analysis Facility

Power and Cooling for Computing at SLAC

0.0

1.0

2.0

3.0

4.0

5.0

6.0

2008 2009 2010 2011 2012 2013 2014 2015 2016 2017

Meg

aWatt

s of C

ompu

ting

Equi

pmen

t

SLAC Computing - Evolution of Power Needs

Maximum MW (every activity one sigma up)

Expectation MW

Minimum MW (every activity one sigma down)

Infrastructure Capacity

04/20/23 8Western Analysis Facility

Page 9: Western Analysis Facility

Infrastructure: 2013 on

• Proposed (Stanford) Scientific Research Computing Facility

• Modular – up to 8 modules• Up to 3MW payload per module• Ambient air cooled• Cheaper than Sun BlackBoxes• But not free! (~$10 per W capital cost)

04/20/23 9Western Analysis Facility

Page 10: Western Analysis Facility

Concept for a Stanford Research Computing Facility at SLAC (~2013)

04/20/23 10Western Analysis Facility

Page 11: Western Analysis Facility

First Two Modules

04/20/23 11Western Analysis Facility

Page 12: Western Analysis Facility

Module Detail

04/20/23 12Western Analysis Facility

Page 13: Western Analysis Facility

Green Savings

04/20/23 13Western Analysis Facility

Page 14: Western Analysis Facility

Ideas on Hosting Costs (1)

Model the acquisition costs, power bill + infrastructure maintenance + support labor for:

1.CPU boxes:• Acquisition

• Annual Costs

Raw Box (Dell R410, 8*2.93 GHz, 24GB, 500GB)

$3776

Network (data + management), rack, cables $981

Power Bill (including cooling) $475

Power/cooling infrastructure maintenance $176

Labor (sysadmin, ATLAS environment) $440

04/20/23 14Western Analysis Facility

Page 15: Western Analysis Facility

Model the acquisition costs, power bill + infrastructure maintenance + support labor for:

2.Disk Space:• Acquisition

• Annual Costs

Raw Box (Sun “Thor”, 2*6-core, 48*1TB = 33TB usable)

$25,400

Network (data + management), rack, cables $3,500

Power Bill (including cooling) $1,381

Power/cooling infrastructure maintenance $512

Labor (sysadmin, ATLAS environment) $4,000

Ideas on Hosting Costs (2)

04/20/23 15Western Analysis Facility

Page 16: Western Analysis Facility

Hosting Terms

The historical approach:

1. Rich Uncle:– University buys or pays for some equipment – SLAC provides space/power/cooling/network/support for free.

Possible future scenarios:

1. Cash economy: – University buys or pays for some equipment and pays for the rack/network/cable

acquisition costs;– University pays the annual operating costs to SLAC

2. Barter economy: – University buys or pays for some equipment– SLAC assigns some fraction of this equipment to meet SLAC HEP program needs– In exchange, SLAC installs, powers, cools and supports the equipment for 3 or 4

years

04/20/23 16Western Analysis Facility

Page 17: Western Analysis Facility

SLAC Strengths

1. Pushing the envelope of Data Intensive Computinge.g. Scalla/xrootd (in use at the SLAC T2)

2. Design and implementation of efficient and scalable computing systems (1000s of boxes)

3. Strongly supportive interactions with the university community (and 10 Gbits/s to Internet2).

Plus a successful ongoing computing operation:• Multi-tiered multi-petabyte storage• ~10,000 cores of CPU• Space/power/cooling continuously evolving

04/20/23 17Western Analysis Facility

Page 18: Western Analysis Facility

SLAC Laboratory Goals

1. Maintain, strengthen and exploit the Core Competency in Data Intensive Computing;

2. Collaborate with universities exploiting the complementary strengths of universities and SLAC.

04/20/23 18Western Analysis Facility

Page 19: Western Analysis Facility

ATLAS Western Analysis FacilityConcept

1. Focus on data-intensive analysis on a “major-HEP-computing-center” scale;

2. Flexible and Nimble to meet the challenge of rapidly evolving analysis needs;

3. Flexible and Nimble to meet the challenge of evolving technologies:• Particular focus on the most effective role for solid state storage

(together with enhancements to data-access software);

4. Close collaboration with US ATLAS university groups:• Make best possible use of SLAC-based and university-based

facilities.

5. Coordinate with ATLAS Analysis Support Centers.

04/20/23 19Western Analysis Facility

Page 20: Western Analysis Facility

ATLAS Western Analysis FacilityPossible Timeline

1. Today:1. Interactive access to SLAC computing available to US ATLAS2. Jobs may be submitted to (most of) the HEP funded CPUs3. Hosting possibilities for Tier 3 equipment

2. Today through 2010: Tests of various types of solid-state storage in various ATLAS roles (conditions DB, TAG access, PROOF-based analysis, xrootd access to AOD …). Collaborate with BNL and ANL.

3. 2010: Re-evaluation of ATLAS analysis needs after experience with real data.

4. 2011 on: WAF implementation as part of overall US ATLAS strategy.

04/20/23 20Western Analysis Facility

Page 21: Western Analysis Facility

Longer Term Goals

Analysis

Simulation

Production

US ATLAS Tier 1 + Tier 2 CPU 2009

AnalysisSimulation

Production

CDF CPU, Fermilab Site 2008

• There is a likelihood that US ATLAS will need additional data-intensive analysis capability:

• There is an even higher likelihood that there will be major software and architectural challenges in ATLAS data analysis.

• SLAC’s goal is to address both issues as part membaer of of US ATLAS.

04/20/23 21Western Analysis Facility


Recommended