Date post: | 19-Dec-2015 |
Category: |
Documents |
View: | 216 times |
Download: | 0 times |
12 July 2004 - 1
Experience Deploying the Experience Deploying the LLarge arge Hadron Collider Hadron Collider CComputing omputing GGridrid
Markus SchulzCERN IT GD-GIS
12 July 2004 - 2
Overview Overview
LHC Computing Grid CERN Challenges for LCG The project Deployment and Status LCG-EGEE Summary
Large Hadron Collider
12 July 2004 - 3
CERN CERN
• CERN European Organization for Nuclear Research• located on the Swiss - French border close to Geneva • funded by 20 european member states
• In addition several observer states and non member states participating in the experiment program
• world’s largest center for particle physics research• provides infrastructure and tools (accelerators etc.)• ~3000 employees and several thousand visiting scientists• 50 years of history (several Nobel Prizes)• The place where the WWW was born (1990) • Next milestone: the LHC (Large Hadron Collider) • http://public.web.cern.ch/public/about/aboutCERN.html
QuickTime™ and aTIFF (Uncompressed) decompressor
are needed to see this picture.
QuickTime™ and aTIFF (Uncompressed) decompressor
are needed to see this picture.QuickTime™ and a
TIFF (Uncompressed) decompressorare needed to see this picture.
12 July 2004 - 4
Challenges for the Challenges for the LLHC HC CComputing omputing GGrid rid http://lcg.web.cern.ch/lcg
• LHC (Large Hadron Collider)• with 27 km of magnets the largest superconducting installation• proton beams collide at an energy of 14TeV • 40 million events per second from each of the 4 experiments• after triggers and filters 100-1000MBytes/second remain• every year ~15PetaByte of data will be recorded• this data has to be reconstructed and analyzed by the users• in addition large computational effort to produce Monte Carlo data
12 July 2004 - 5
The Four Experiments at LHCThe Four Experiments at LHC
LHCbFederico.carminati , EU review presentation
12 July 2004 - 6
15 PetaByte/year15 PetaByte/year
• 15 PetaByte/year have to be:• Recorded• Cataloged• Managed • Distributed• Processed
50 CD-ROM
= 35 GB
6 cm
Concorde(15 Km)
Balloon(30 Km)
CD stack with1 year LHC data!(~ 20 Km)
Mt. Blanc(4.8 Km)
12 July 2004 - 7
Core TasksCore Tasks
• Reconstruction: transform signals from the detector to physical properties
– energy, charge, tracks, momentum, particle id.– this task is computational intensive and has modest I/O requirements– structured activity (production manager)
• Simulation: start from the theory and compute the responds of the detector
– very computational intensive– structured activity, but larger number of parallel activities
• Analysis: complex algorithms, search for similar structures to extract physics
– very I/O intensive, large number of files involved– access to data cannot be effectively coordinated – iterative, parallel activities of hundreds of physicists
12 July 2004 - 8
Computing NeedsComputing Needs
• Some 100 Million SPECInt2000 are needed• A 3 GHz Pentium IV ~ 1K SPECInt2000• O(100k) CPUs are needed
12 July 2004 - 9
Large and distributed user communityLarge and distributed user community
• CERN Collaborators• > 6000 users from 450 institutes• none has all the required computing• all have access to some computing
Europe: 267 institutes4603 users
Elsewhere: 208 institutes1632 users
Solution: Connect all the resources into a computing grid
12 July 2004 - 10
The LCG Project The LCG Project (and what it isn’t)(and what it isn’t)
• Mission• To prepare, deploy and operate the computing environment for the
experiments to analyze the data from the LHC detectors
• Two phases:• Phase 1: 2002 – 2005• Build a prototype, based on existing grid middleware• Deploy and run a production service• Produce the Technical Design Report for the final system
• Phase 2: 2006 – 2008 • Build and commission the initial LHC computing environment
LCG is NOT a development project for middlewarebut problem fixing is permitted (even if writing code is required)
12 July 2004 - 11
LCG Time LineLCG Time Line
Testing, with simulated event productions
2003
2004
2005
2006
2007
first data
physicscomputing service
open LCG-1 – 15 Sept
* TDR – technical design report
Computing models
TDR for the Phase 2 grid
experiment setup & preparation
Phase 2 service in production
Phase 2 service acquisition, installation, commissioning
principal service for LHC data challenges
LCG-3 – second generation middlewarevalidation of computing models
Second generation middleware prototyping, development
LCG-2 - upgraded middleware, mgt. and ops tools
12 July 2004 - 12
RAL
IN2P3
BNL
FZK
CNAF
PIC ICEPP
FNAL
USCNIKHEFKrakow
CIEMAT
Rome
Taipei
TRIUMF
CSCS
Legnaro
UB
IFCA
IC
MSU
Prague
Budapest
Cambridge
Tier-1small
centres Tier-2desktopsportables
LCG Scale and Computing ModelLCG Scale and Computing Model
• Tier-0• reconstruct (ESD)• record raw and ESD• distribute data to tier-1
• Tier-1• data heavy analysis• permanent, managed grid-enabled
storage (raw, analysis, ESD), MSS• reprocessing• regional support
• Tier-2• managed disk storage• simulation• end user analysis• parallel interactive analysis
Processing M SI2000**
Disk PetaBytes
Mass Storage
PetaBytes
CERN 20 5 20
Major data handling centres
(Tier 1)45 20 18
Other large centres (Tier 2)
40 12 5
Totals 105 37 43
** Current fast processor ~1K SI2000
Current estimates of Computing Resources needed at Major LHC Centres
First full year of data - 2008
Data distribution~70Gbits/sec
12 July 2004 - 13
LCG - a CollaborationLCG - a Collaboration
• Building and operating the LHC Grid - a collaboration between
• The physicists and computing specialists from the experiments
• The projects in Europe and the US that have been developing Grid middleware
• European DataGrid (EDG)• US Virtual Data Toolkit (Globus, Condor, PPDG,
iVDGL, GriPhyN)
• The regional and national computing centres that provide resources for LHC
• some contribution from HP (tier2 centers)
• The research networks
Researchers
Software Engineers
Service Providers
12 July 2004 - 14
LCG-2 SoftwareLCG-2 Software
• LCG-2_1_0 core packages:• VDT (Globus2)• EDG WP1 (Resource Broker)• EDG WP2 (Replica Management tools)
• One central RMC and LRC for each VO, located at CERN, ORACLE backend
• Several bits from other WPs (Config objects, InfoProviders, Packaging…)
• GLUE 1.1 (Information schema) + few essential LCG extensions• MDS based Information System with LCG enhancements
• Almost all components have gone through some reengineering
• robustness• scalability• efficiency • adaptation to local fabrics
12 July 2004 - 15
LCG-2 SoftwareLCG-2 Software
• Authentication and Authorization• Globus GSI based on X509 certificates• LCG established trust relationship between the CAs in the project• Virtual Organizations (VOs) registration hosted at different sites
• Data Management Tools• Catalogues keep track of replicas (Replica Metadata Catalog, Local
Replica Catalog) • SRM interface for several HMSS and disk pools• Wide area transport via GridFTP
RLI
LFN-1
LFN-2
LFN-3
GUID
RMC
SURL-1
LRC
SURL-2
lfn:Toto.ps Not4Humans srm://host.domain/path/file
12 July 2004 - 16
LCG-2 SoftwareLCG-2 Software
• Information System• Globus MDS based for information gathering on a site• LDAP + lightweight DB based system for collecting data from sites
• LCG-BDII solved scalability problem of Globus2 MDS (>200sites tested)• Contains information on capacity, capability, utilization and state of
services (computing, storage, cataloges..)
• Work Load Data Management Tools• Matches user requests with the resources available for a VO
• requirements formulated in JDL (classadds)• user tunable ranking of resources
• Uses RLS and information system • Keeps state of jobs and manages extension of credentials
• input output sandboxes • proxy renewal….
• Interfaces to local batch systems via a gateway node
12 July 2004 - 17
LCG-2 SoftwareLCG-2 Software
Job Status
submitted
waiting
ready
scheduled
running
done
cleared
UIReplicaCatalog
Inform.Service
NetworkServer
Job Contr.-
CondorG
WorkloadManager
RB node
CE characts& status
SE characts& status
RBstorage
Match-Maker/ Broker
JobAdapter
Log Monitor
Logging &Bookkeeping
sandbox
Matching
Job Adapter
On CE
Processed
Output back
User done
Arrived on RBInput Sandbox is what you take with you to the node Output Sandbox is what you get back
Failed jobs are resubmitted
12 July 2004 - 18
LCG Grid Deployment AreaLCG Grid Deployment Area Goal: - deploy & operate a prototype LHC computing
environment Scope:
• Integrate a set of middleware and coordinate and support its deployment to the regional centres
• Provide operational services to enable running as a production-quality service
• Provide assistance to the experiments in integrating their software and deploying in LCG; Provide direct user support
Deployment Goals for LCG-2• Production service for Data Challenges in 2004
• Initially focused on batch production work • Experience in close collaboration between the Regional Centres• Learn how to maintain and operate a global grid• Focus on building a production-quality service
• Focus on robustness, fault-tolerance, predictability, and supportability• Understand how LCG can be integrated into the sites’ physics
computing services
12 July 2004 - 19
DeploymentArea ManagerDeployment
Area ManagerGrid Deployment
BoardGrid Deployment
Board
CertificationTeam
CertificationTeam
DeploymentTeam
DeploymentTeam
ExperimentIntegration
Team
ExperimentIntegration
Team
Testing groupTesting group
Security group
Security group
Storage group
Storage group
GDB task forces
JTB
HEPiX
GGF
Grid Projects:EDG,Trillium,Grid3,etc
Regional Centres
LHC Experiments
LCG Deployment Area
LCG Deployment Organisation and Collaborations
OperationsCentres- RAL
OperationsCentres- RAL
Call Centres- FZK
Call Centres- FZK
Advises, informs,Sets policy
Set requirements
Set requirements
Col
labo
rativ
e ac
tiviti
es
participate
participate
12 July 2004 - 20
ImplementationImplementation A core team at CERN – Grid Deployment group (~30) Collaboration of the regional centres – through the Grid Deployment
Board Partners take responsibility for specific tasks (e.g. GOCs, GUS) Focussed task forces as needed Collaborative joint projects – via JTB, grid projects, etc.
• CERN deployment group • Core preparation, certification, deployment, and support activities• Integration, packaging, debugging, development of missing tools,• Deployment coordination & support, security & VO management,• Experiment integration and support
• GDB: Country representatives for regional centres• Address policy, operational issues that require general agreement• Brokered agreements on:
• Initial shape of LCG-1 via 5 working groups• Security• What is deployed
12 July 2004 - 21
Operations Services Operations Services
• Operations Service:• RAL (UK) is leading sub-project on developing operations services• Initial prototype http://www.grid-support.ac.uk/GOC/
• Basic monitoring tools• Mail lists for problem resolution• GDB database containing contact and site information• Working on defining policies for operation, responsibilities (draft document)• Working on grid wide accounting
• Monitoring:• GridICE (development of DataTag Nagios-based tools) • GridPP job submission monitoring
Many more like http://goc.grid.sinica.edu.tw/gstat/
• Deployment and operation support:• Hierarchical model
• CERN acts as 1st level support for the Tier 1 centres• Tier 1 centres provide 1st level support for associated Tier 2s
12 July 2004 - 22
User Support User Support
• Central model for user support• VOs provide 1st level triage• FZK (germany) leading sub-project to develop user support services • Web portal for problem reporting http://gus.fzk.de/• Experiments contacts send problems through the FZK portal• During the data challenges the experiments used a direct channel
via the GD teams.
• Experiment integration support by CERN based group
• close collaboration during data challenges
• Documentation• Installation guides (manual and management tool based)
• See:http://grid-deployment.web.cern.ch/grid-deployment/cgi-bin/index.cgi?var=releases
• Rather comprehensive user guides
12 July 2004 - 23
SecuritySecurity
• LCG Security Group • LCG1 usage rules (still used by LCG2) • Registration procedures and VO management
• Agreement to collect only minimal amount of personal data
• Initial audit requirements are defined• Initial incident response procedures
• Site security contacts etc. are defined
• Set of trusted CAs (including Fermilab online KCA)• Draft of security policy (to be finished by end of year)• Web site http://proj-lcg-security.web.cern.ch/proj
-lcg-security/
12 July 2004 - 24
Certification and release cycles Certification and release cycles
Developers
CERTIFICATIONTESTING
CERTIFICATIONTESTING SERVICESSERVICES
Integrate
BasicFunctionality
Tests
Run testsC&T suitesSite suites
RunCertification
Matrix
Releasecandidate
tag
PR
E-P
RO
DU
CT
ION
PR
OD
UC
TIO
N
APPINTEGR
APPINTEGR
Certifiedrelease
tag
DE
VE
LO
PM
EN
T &
IN
TE
GR
AT
ION
UN
IT &
FU
NC
TIO
NA
L T
ES
TIN
G
DevTag
HEPEXPTS
BIO-MED
OTHERTBD
APPSSW
Installation
DE
PL
OY
ME
NT
PR
EP
AR
AT
ION
Deploymentrelease
tag
DEPLOYDEPLOY
Productiontag
12 July 2004 - 25
2003-2004 Milestones2003-2004 Milestones
Project Level 1 Deployment milestones for 2003: July: Introduce the initial publicly available LCG-1 global grid service
• With 10 Tier 1 centres in 3 continents November: Expanded LCG-1 service with resources and functionality
sufficient for the 2004 Computing Data Challenges• Additional Tier 1 centres, several Tier 2 centres – more countries• Expanded resources at Tier 1s • Agreed performance and reliability targets
• 2004 the “LHC Data Challenges”“LHC Data Challenges”• Large-scale tests of the experiments’ computing models, processing
chains, grid technology readiness, operating infrastructure• ALICE and CMS data challenges started at the beginning of March• LHCb and ATLAS – started in May/June• The big challenge for this year - data –
- file catalogue, (million of files) - replica management, - database access, - integrating all available mass storage systems
(several hundred TByte)
12 July 2004 - 26
HistoryHistory
• Jan 2003 GDB agreed to take VDT and EDG components• March 2003 LCG-0
• existing middleware, waiting for EDG-2 release
• September 2003 LCG-1 • 3 month late -> reduced functionality• extensive certification process• improved stability (RB, Information system)• integrated 32 sites ~300 CPUs• operated until early January, first use for production
• December 2003 LCG-2 • Full set of functionality for DCs, but only “classic SE”, first MSS integration• Deployed in January, • Data challenges started in February -> testing in production• Large sites integrate resources into LCG (MSS and farms)
• Mai 2004 -> now releases with incrementaly:• Improved services • SRM enabled storage for disk and MSS systems
12 July 2004 - 27
LCG1 Experience (2003)LCG1 Experience (2003)
• Integrate sites and operate a grid• Problems: only 60% of personnel, software was late• introduced hierarchical support model (primary and secondary sites)
• worked well for some regions (less for others)• installation and configuration was an issue
• only time to package software for the LCFGng tool (problematic)• not sufficient documentation (partially compensated by travel)• manual installation procedure documented when new staff arrived
• communication/cooperation between sites needed to be established• deployed MDS + EDG-BDII in a robust way
• redundant regional GIISes • vastly improved the scalability and robustness of the information system
• upgrades, especially non backward compatible ones took very long• not all sites showed the same dedication
• still some problems with the reliability of some of the core services
Big step forward
12 July 2004 - 28
LCG2LCG2
• Operate large scale productionproduction service• Started with 8 “core” sites
• each bringing significant resources • sufficient experience to react quickly• weekly core site phone conference• weekly meeting with each of the experiments• weekly joined meeting of the sites and the experiments (GDA)
• Introduced a testZone for new sites• LCG1 showed that ill configured sites can affect all sites• sites stay in the testZone until they have been stable for some time
• Further improved (simplified) information system• addresses: manageability, improves: robustness and scalability• allows partitioning of the grid into independent views
• Introduced local testbed for experiment integration• rapid feedback on functionality from the experiments• triggered several changes to the RLS system
12 July 2004 - 29
LCG2 LCG2
• Focus on integrating local resources• batch systems at CNAF, CERN, NIKHEF already integrated• MSS systems with CASTOR at several sites, enstore at FNAL
• Experiment software distribution mechanism• based on shared file system with access for privileged users• tool to publish the installed software in the information system• needs to be as transparent as possible. (Some work to be done)
• Improved documentation and installation • sites have the choice to use LCFGng or follow a manual installation guide• generic config. description eases integration with local tools • documentation includes simple tests, sites join in a better state• improved readability by going to HTML and PDF
• Release page
12 July 2004 - 30
Site Reports GIIS Name Site Reports GIIS NameCAVN farm012.hep.phy.cam.ac.uk GLASGOW ce1-gla.scotgrid.ac.ukCERN-LCG2 lxn1181.cern.ch GSI lcg03.gsi.deCIEMAT lcg02.ciemat.es HEPHYUIBK grid.uibk.ac.atCNAF-LCG2 wn-04-07-02-a.cr.cnaf.infn.itHP-PUERTO-RICO ce.prd.hp.comFNAL-LCG2 hotdog46.fnal.gov INFN-NAPOLI atlasce01.na.infn.itFZK-LCG2 gridkap01.fzk.de INFN-ROMA1 t2-ce-01.roma1.infn.itIFCA gtbcg12.ifca.unican.es IPAS testbed001.phys.sinica.edu.twIFIC lcg2ce.ific.uv.es JINR lxt03.jinr.ruIMPERIAL gw39.hep.ph.ic.ac.uk Karlruhe
ekp-lcg-ce.physik.uni-karlsruhe.de
INFN-LEGNARO t2-ce-01.lnl.infn.it LANC lunegw.lancs.ac.ukINFN-MILANO t2-ce-01.mi.infn.it LIP ce01.lip.ptINFN-TORINO grid008.to.infn.it MANCHESTER bohr0001.tier2.hep.man.ac.ukKRAKOW zeus02.cyf-kr.edu.pl MONTREAL lcg-ce.lps.umontreal.caNCU-TAIWAN grid01.phy.ncu.edu.tw NCP pcncp04.ncp.edu.pknikhef.nl tbn18.nikhef.nl PRAGUE golias25.farm.particle.czPIC-LCG2 lcgce02.ifae.es PRAGUE-CESNET skurut17.cesnet.czRAL-LCG2 lcgce02.gridpp.rl.ac.uk QMUL ce01.ph.qmul.ac.ukTaiwan-LCG2 lcg00125.grid.sinica.edu.tw RALPP heplnx131.pp.rl.ac.ukTRIUMF lcgce01.triumf.ca SARA mu6.matrix.sara.nlUSC-LCG2 lcg-ce.usc.cesga.es SHEF ce.gridpp.shef.ac.ukWUPPERTAL grid-ce.physik.uni-wuppertal.deSINP lcg06.sinp.msu.ru#HG-01-GRNET ce01.isabella.grnet.gr TAU lcfgng.cs.tau.ac.ilAACHEN lcg2-CE.physik.rwth-aachen.deTOKYO dgce0.icepp.jpALBERTA lcgce01.nic.ualberta.ca TORONTO bigmac-lcg-ce.physics.utoronto.caBHAM-HEP epcf36.ph.bham.ac.uk TRIUMF-GC lcgce02.triumf.caBNL atlasgrid04.usatlas.bnl.gov UAM grid003.ft.uam.esBUDAPEST grid109.kfki.hu UB-BARCELONA lcg-ce.ecm.ub.esCARLETON lcg02.physics.carleton.ca UCL pc31.hep.ucl.ac.ukCCIN2P3-LCG2 cclcgceli01.in2p3.fr UCL-CCC ce-a.ccc.ucl.ac.ukCLERMONT clrglop195.in2p3.fr WEIZMANN wipp-ce.weizmann.ac.ilCSCS compute-0-10.cscs.ch ITEP ceitep.itep.ruDESY grid-ce1.desy.de FRASCATI atlasce.lnf.infn.it
12 July 2004 - 31
Sites in LCG2 July 8 2004Sites in LCG2 July 8 2004
• 22 Countries• 63 Sites
• 49 Europe, 2 US, 5 Canada, 6 Asia, 1 HP• Coming: New Zealand, China, Korea
other HP (Brazil, Singapore)• 6100 cpu
12 July 2004 - 32
UsageUsage
• Hard to measure:• VOs “Pick” services and add
own components for• job submission, file catalogs,
replication …
• we have no central control of the resources
• accounting has to be improved
• File catalogues (only used by 2 VOs )
• ~ 2.5 Million entries
QuickTime™ and aTIFF (Uncompressed) decompressor
are needed to see this picture.
12 July 2004 - 35
2004 Data Challenges2004 Data Challenges
• LHC experiments use multiple grids and additional services• Integration, Interoperability• expect some central resource negotiation concerning:
– queue length, memory,scratch space, storage etc.
• Service provision• planned to provide shared services RBs, BDIIs, UIs etc• experiments need to augment the services on the UI and need to define their
super/subsets of the grid– individual RB/BDII/UIs for each VO (optional on one node)
• Scalable services for data transport needed• DNS switched access to gridFTP
• Performance issues with several tools (RLS, Info System, RBs) • most understood, work around and fixes implemented and part of 2_1_0
• Local Resource Managers (batch systems) are too smart for Glue• GlueSchema can’t express the richness of batch systems (LSF etc.)
– load distribution not understandable for users (looks wrong)– problem is understood and workaround in preperation
12 July 2004 - 36
Interoperability Interoperability
• Several grid infrastructures for LHC experiments:• LCG-2/EGEE, Grid2003/OSG, NorduGrid, other national grids
• LCG/EGEE explicit goals to interoperate• One of LCG service challenges• Joint projects on storage elements, file catalogues, VO management, etc.
• Most are VDT (or at least Globus-based)• Grid2003 & LCG use GLUE schema
• Issues are:• File catalogues, information schema, etc at technical level• Policy and semantic issues
12 July 2004 - 37
Developments in 2004Developments in 2004
• General:• LCG-2 will be the service run in 2004 – aim to evolve incrementally• Goal is to run a stable service• Service challenges (data transport (500 MB/s one week), jobs, interoperability)
• Some functional improvements:• Extend access to MSS – tape systems, and managed disk pools• Distributed vs replicated replica catalogs – with Oracle back-ends
• To avoid reliance on single service instances
• Operational improvements:• Monitoring systems – move towards proactive problem finding, ability to take
sites on/offline; experiment monitoring (R-GMA), accounting• Control system• “Cookbook” to cover planning, installation and operation• Activate regional centres to provide and support services
• this has improved over time, but in general there is too little sharing of tasks
• Address integration issues:• With large clusters (on non routed networks), with storage systems, with different
OSs, • Integrating with other experiments and apps
12 July 2004 - 38
Changing landscapeChanging landscape
• The view of grid environments has changed in the past year• From
• A view where all LHC sites would run a consistent and identical set of middleware,
• To• A view where large sites must support many experiments each of which
have grid requirements• National grid infrastructures are coming – catering to many applications,
and not necessarily driven by HEP requirements
• We have to focus on interoperating between potentially diverse infrastructures (“grid federations”)
• At the moment these have underlying same m/w• But modes of use and policies are different (IS, file catalogues,..)
• Need to have agreed services, interfaces, protocols• The situation is now more complex than anticipated
12/05 2004 - 39
LCG – EGEE
• LCG-2 will be the production service during 2004• Will also form basis of EGEE initial production service• Will be maintained as a stable service• Will continue to be developed
• Expect in parallel a development service – Q204• Based on EGEE middleware prototypes• Run as a service on a subset of EGEE/LCG production sites
• The core infrastructure of the LCG and EGEE grids will be operated as a single service
• LCG includes US and Asia, EGEE includes other sciences
• The ROCs support Resource Centres and applications • Similar to LCG primary sites • Some ROCs and LCG primary sites will be merged
• LCG Deployment Manager will be the EGEE Operations Manager • Will be member of PEB of both projects
EGEE
LCG
geographical
appl
icat
ions
12/05 2004 - 41
Summary
• 2003 – first MC event production
• In 2004 we must show that we can handle the data – supporting simple computing models
-- This is the key goal of the 2004 Data Challenges
• Target for end of this year –• Basic model demonstrated
using current grid middleware
• All Tier-1s and ~25% of Tier-2soperating a reliable service
• Validate security model, understand storage model
• Clear idea of the performance, scaling, operations and
management issues
2004
2005
2006
2007
first data
Initial service in operation
Decisions on final core middleware
Demonstrate core data handling and batch analysis
Installation and commissioning
12/05 2004 - 42
Summary II
• Getting the grid services going in time for LHC will be even harder than we think today
• The service for LHC must be in permanent operation by September 2006 - CERN↔Tier-1s ↔ major Tier-2s
• So we will spend the first part of 2006 in installation and commissioning
the technology we use must be working (really working) by the end of 2006
a year from now we will have to decide which middleware we are going to use
• From now until the end of 2006 we have to turn prototypes into pilot services
2004
2005
2006
2007
first data
Initial service in operation
Decisions on final core middleware
Demonstrate core data handling and batch analysis
Installation and commissioning
Prototype services pilot services
12/05 2004 - 43
Conclusion (the last “last” slide)
• There are still many questions about grids & data handling
• EGEE provides LCG with opportunities - • to develop an operational grid in an
international multi-science context• to influence the evolution of a generic
middleware package• But the LHC clock is ticking - deadlines will
dictate simplicity and pragmatism• LCG has long-term requirements
– and at present EGEE is a two-year project
• LCG must encompass non-European resources and grids (based on different technologies)
• No shortage of challenges and opportunities
2004
2005
2006
2007
first data
Initial service in operation
Decisions on final core middleware
Demonstrate core data handling and batch analysis
Installation and commissioning