+ All Categories
Home > Documents > US Grid Efforts

US Grid Efforts

Date post: 17-Jan-2016
Category:
Upload: lynton
View: 37 times
Download: 1 times
Share this document with a friend
Description:
US Grid Efforts. Lee Lueking D0 Remote Analysis Workshop February 12, 2002. - PowerPoint PPT Presentation
Popular Tags:
30
US Grid Efforts US Grid Efforts Lee Lueking D0 Remote Analysis Workshop February 12, 2002
Transcript
Page 1: US Grid Efforts

US Grid EffortsUS Grid Efforts

Lee Lueking

D0 Remote Analysis Workshop

February 12, 2002

Page 2: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 2

All of these projects are working towards the common goal of providing transparent access to the massively distributed computing infrastructure that is needed to meet the challenges of modern experiments … (From the EU DataTAG proposal)

Page 3: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 3

Grid Projects TimelineGrid Projects Timeline

Q3 00

Q4 00

Q4 01

Q3 01

Q2 01

Q1 01

Q1 02

GriPhyN: $11.9M+$1.6M

PPDG:$9.5M

iVDGL:$13.65M

EU DataGrid: $9.3M

EU DataTAG:4M Euros

GridPP:

Page 4: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 4

PPDG PPDG

Develop, acquire and deliver vitally needed Grid-enabled tools for data-intensive requirements of particle and nuclear physics.

Collaboration of computer scientists with a strong record in distributed computing and Grid technology, and physicists with leading roles in the software and network infrastructures for major high-energy and nuclear experiments.

Goals and plans are ultimately guided by the immediate, medium-term and longer-term needs and perspectives of the physics experiments.

Page 5: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 5

GriPhyN: Grid Physics NetworkGriPhyN: Grid Physics Network

Virtual data technologies. Advances are required in information models and in new methods of cataloging, characterizing, validating, and archiving software components to implement virtual data manipulations

Policy-driven request planning and scheduling of networked data and computational resources. We require mechanisms for representing and enforcing both local and global policy constraints and new policy-aware resource discovery techniques.

Management of transactions and task-execution across national-scale and worldwide virtual organizations. New mechanisms are needed to meet user requirements for performance, reliability, and cost.

Page 6: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 6

iVDGL:International Virtual iVDGL:International Virtual Data Grid LaboratoryData Grid Laboratory

The iVDGL will provide a global computing resource for several leading international experiments in physics and astronomy,

Global services and centralized monitoring, management, and support functions functions will be coordinated by the Grid Operations Center (GOC) located at Indiana University, with technical effort provided by GOC staff, iVDGL site staff, and the CS support teams.

GriPhyN and Particle Physics Data Grid will provide the basic R&D and software toolkits needed for the laboratory.

The European Union DataGrid is also a major participant and will contribute basic technologies and tools.

The iVDGL will be based on the open Grid infrastructure provided by the Globus Toolkit and will also build on other technologies such as Condor resource management tools.

Page 7: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 7

Comparison of PPDG and iVDGLComparison of PPDG and iVDGLPPDG iVDGL

Funding US DOE approved

1/1/3/3/3 $M, 99 – 03

US NSF proposed

3/3/3/3/3 $M, 02 – 06

Computer Science

Globus (Foster), Condor (Livny), SDM (Shoshani), SRB (Moore)

Globus (Foster, Kesselman), Condor (Livny)

Physics BaBar, Dzero, STAR, JLAB, ATLAS, CMS

ATLAS, CMS, LIGO,

SDSS, NVO

National Laboratories

BNL, Fermilab, JLAB, SLAC, ANL, LBNL

ANL,BNL, Fermilab (all unfunded collaborators)

Universities Caltech, SDSS, UCSD, Wisconsin

Florida, Chicago, Caltech, UCSD, Indiana, Boston, Wisconsin at Milwaukee, Pennsylvania State, Johns Hopkins, Wisconsin at Madison, Northwestern, USC, UT Brownsville, Hampton, Salish Kootenai College

Hardware None ~20% of funding(Tier-2 Centers)

Network No funding requested No funding requestedDataTAG complementary

Page 8: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 8

PPDG CollaboratorsPPDG Collaborators

Page 9: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 9

PPDG Computer Science GroupsPPDG Computer Science GroupsCondor – develop, implement, deploy, and evaluate mechanisms and policies

that support High Throughput Computing on large collections of computing resources with distributed ownership.

http://www.cs.wisc.edu/condor/

Globus - developing fundamental technologies needed to build persistent environments that enable software applications to integrate instruments, displays, computational and information resources that are managed by diverse organizations in widespread locations

http://www.globus.org/

SDM - Scientific Data Management Research Group – optimized and standardized access to storage systems

http://gizmo.lbl.gov/DM.html

Storage Resource Broker - client-server middleware that provides a uniform interface for connecting to heterogeneous data resources over a network and cataloging/accessing replicated data sets.

http://www.npaci.edu/DICE/SRB/index.html

Page 10: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 10

Delivery of End-to-End ApplicationsDelivery of End-to-End Applications& Integrated Production Systems & Integrated Production Systems

to allow thousands of physicists to share data & computing resources for scientific processing and analyses

Operators & Users

Resources: Computers, Storage, Networks

PPDG Focus:

- Robust Data Replication

- Intelligent Job Placement and Scheduling

- Management of Storage Resources

- Monitoring and Information of Global Services

Relies on Grid infrastructure:- Security & Policy- High Speed Data Transfer- Network management

Page 11: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 11

Common ServicesCommon Services Job Description Language Scheduling and Management of Processing and Data

Placement Activities Monitoring and Status Reporting Storage Resource Management Reliable Replica Management Services File Transfer Services Collect and Document Current Experimental Practices R & D, Evaluation Authentication, Authorization, and Security End-to-End Applications and Testbeds

Page 12: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 12

Project Activities,Project Activities,End-to-End ApplicationsEnd-to-End Applications

and Cross-Cut Pilotsand Cross-Cut Pilots

Project Activities are focused Experiment – Computer Science Collaborative developments.

Replicated data sets for science analysis – BaBar, CMS, STARDistributed Monte Carlo production services – ATLAS, D0, CMSCommon storage management and interfaces – STAR, JLAB

End-to-End Applications used in Experiment data handling systems to give real-world requirements, testing and feedback.

Error reporting and responseFault tolerant integration of complex components

Cross-Cut Pilots for common services and policies Certificate Authority policy and authenticationFile transfer standards and protocolsResource Monitoring – networks, computers, storage.

Page 13: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 13

Super Computing 2001 in Denver

Page 14: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 14

Page 15: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 15

PPDG activities as part of the PPDG activities as part of the Global Grid CommunityGlobal Grid Community

Coordination with other Grid Projects in our field:GriPhyN – Grid for Physics NetworkEuropean DataGridStorage Resource Management collaboratoryHENP Data Grid Coordination Committee

Participation in Experiment and Grid deployments in our field:ATLAS, BaBar, CMS, D0, Star, JLAB experiment data handling systemsiVDGL/DataTAG – International Virtual Data Grid LaboratoryUse DTF computational facilities?

Active in Standards Committees:Internet2 HENP Working Group Global Grid Forum

Page 16: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 16

PPDG and GridPP ProjectsPPDG and GridPP Projects

Use of Standard Middleware to Promote Interoperability Move to Globus infrastructure: GSI, GridFTP Use of Condor as a supported system for job

submission Publish availability of resources and file catalog

Additional Grid Functionality for Job Specification, Submission, and Tracking Use Condor for migration and check pointing Enhanced job specification language and services

Enhanced Monitoring and Diagnostic Capabilities Fabric Management

Page 17: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 17

PPDG Management and CoordinationPPDG Management and CoordinationPIs

Livny, Newman, Mount

Steering CommitteeRuth Pordes, Chair

Doug Olson, Physics Deputy ChairMiron Livny, Computer Science Deputy Chair

Computer Science Group RepresentativesPhysics Experiment Representatives

PIs (ex officio)

STAR SDM BaBar SRB JLAB ATLAS Globus CMS Condor DZero

Executive Team (>1.0 FTE on PPDG)Steering Committee Chair

Steering Committee Physics and CS Deputy Chairs

Page 18: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 18

iVDGLiVDGL

International Virtual-Data Grid Laboratory A global Grid laboratory with participation from US, EU, Asia,

etc. A place to conduct Data Grid tests “at scale” A mechanism to create common Grid infrastructure A facility to perform production exercises for LHC experiments A laboratory for other disciplines to perform Data Grid tests

“We propose to create, operate and evaluate, over asustained period of time, an international researchlaboratory for data-intensive science.”

From NSF proposal, 2001

Page 19: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 19

iVDGL Summary InformationiVDGL Summary Information Principal components (as seen by USA)

Tier1 sites (laboratories) Tier2 sites (universities and other institutes) Selected Tier3 sites (universities) Fast networks: US, Europe, transatlantic International Grid Operations Center (iGOC) Computer Science support teams Coordination, management

Proposed international partners Initially US, EU, Japan, Australia Other world regions later Discussions w/ Russia, China, Pakistan, India, South America

Complementary EU project: DataTAG Transatlantic network from CERN to STAR-TAP (+ people) Initially 2.5 Gb/s

Page 20: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 20

US Proposal to NSFUS Proposal to NSF US proposal approved by NSF Sept. 25, 2001

“Part 2” of GriPhyN project Much more application oriented than first GriPhyN proposal $15M, 5 years @ $3M per year (huge constraint) CMS + ATLAS + LIGO + SDSS/NVO + Computer Science

Scope of US proposal Deploy Grid laboratory with international partners Acquire Tier2 hardware, Tier2 support personnel Integrate of Grid software into applications CS support teams (+ 6 UK Fellows) to harden tools Establish International Grid Operations Center (iGOC) Deploy hardware at 3 minority institutions (Tier3)

Page 21: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 21

US iVDGL Proposal ParticipantsUS iVDGL Proposal Participants

T2/Software

CS support

T3/Outreach

T1/Labs

U Florida CMS Caltech CMS, LIGO UC San Diego CMS, CS Indiana U ATLAS, iGOC Boston U ATLAS U Wisconsin, Milwaukee LIGO Penn State LIGO Johns Hopkins SDSS, NVO U Chicago CS U Southern California CS U Wisconsin, Madison CS Salish Kootenai Outreach, LIGO Hampton U Outreach, ATLAS U Texas, Brownsville Outreach, LIGO Fermilab CMS, SDSS, NVO Brookhaven ATLAS Argonne Lab ATLAS, CS

Page 22: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 22

iVDGL Partners iVDGL Partners

National partners PPDG (Particle Physics Data Grid ) DTF: Distributed Terascale Facility CAL-IT2 (new California Grid initiative)

Current international partners EU-DataGrid UK PPARC funding agency UK Core e-Science Program 6 UK Fellowships INFN (Italy) 2 Japanese institutes 1 Australian institute (APAC)

Page 23: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 23

iVDGL Map Circa 2002-2003iVDGL Map Circa 2002-2003

Tier0/1 facility

Tier2 facility

10 Gbps link

2.5 Gbps link

622 Mbps link

Other link

Tier3 facility

Page 24: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 24

iVDGL RequirementsiVDGL Requirements

Realistic scale In number, diversity, distribution, network connectivity

Delegated management and local autonomy Management needed to operate as large, single facility Autonomy needed for sites and experiments

Support large-scale experimentation To provide useful information for building real Data Grids

Robust operation For long running applications in complex environment

Instrumentation and monitoring Required for an experimental facility

Integration with international “cyberinfrastructure” Extensibility

Page 25: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 25

ApproachApproach

Define a laboratory architecture Define expected laboratory functions Build in scalability, extensibility, reproducibility Define instrumentation, monitoring Establish CS support teams (develop/harden tools, support

users) Define working relationship, coordination with partners

Create and operate global-scale laboratory Deploy hardware, software, personnel at Tier2, Tier3 sites Establish iGOC, single point of contact for monitoring,

support, … Help international partners establish sites

Evaluate and improve iVDGL through experimentation CS support teams will work with experiments Extend results to partners

Engage underrepresented groups Integrate minority institutions as Tier3 sites

Page 26: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 26

iVDGL as a LaboratoryiVDGL as a Laboratory

Grid Exercises “Easy”, intra-experiment tests first (10-30%, national,

transatlantic) “Harder” wide-scale tests later (30-100% of all

resources) CMS is already conducting transcontinental simulation

productions Operation as a facility

Common software, central installation to ensure compatibility

CS teams to “harden” tools, support applications iGOC to monitor performance, handle problems

Page 27: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 27

Emphasize Simple OperationEmphasize Simple Operation

“Local” control of resources vitally important (Site level or national level) Experiments, politics demand it

Operate mostly as a “partitioned” testbed (Experiment, nation, etc.) Avoids excessive coordination Allows software tests in different partitions

Hierarchy of operation must be defined E.g., (1) National + experiment, (2) inter-expt., (3)

global tests

Page 28: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 28

Other DisciplinesOther Disciplines

Use by other disciplines Expected to be at the 10% level Other HENP experiments Virtual Observatory (VO) community in Europe/US Gravity wave community in

Europe/US/Australia/Japan Earthquake engineering Bioinformatics Our CS colleagues (wide scale tests)

Page 29: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 29

US iVDGL Management and CoordinationUS iVDGL Management and Coordination

Project DirectorsAvery, Foster

Project Coordination GroupProject Coordinator

Project DirectorsCoordinators of Systems Integration, Education/Outreach

Physics Experiment RepresentativesUniversity Research Center or Group Representatives

PACI Representatives

iVDGL Design and Deployment

Integration with Applications

University Research Centers / Groups

International Grid Operations Center

Collaboration Board (Advisory)

External Advisory Board

Page 30: US Grid Efforts

February 12, 2002 Lee Lueking - D0 RACE 30

ConclusionConclusion

PPDG, and iVDGL are complementary in their approach and deliverables.

These efforts, along with our European partners will provide exciting new ways to share data and computing resources.

Dzero Grid involvement offers many challenges, but even more opportunities.

Acknowledgements: Richard Mount (SLAC), Paul Avery (University of Florida), Ruth Pordes (FNAL).


Recommended