Addendum to the Particle Physics e-Science Programme Proposal
The UK Grid for Particle Physics Collaboration
GridPP
University of Birmingham,
University of Bristol,
Brunel University,
CERN, European Organization for Nuclear Research,
University of Cambridge,
University of Durham,
University of Edinburgh,
University of Glasgow,
Imperial College of Science, Technology and Medicine,
Lancaster University,
University of Liverpool,
University of Manchester,
Oxford University,
Queen Mary, University of London,
Royal Holloway, University of London,
Rutherford Appleton Laboratory,
University of Sheffield,
University of Sussex,
University of Wales Swansea,
University College London.
Contacts
Dr. Tony Doyle –
A.Doyle@physics.gla.ac.ukDr. Steve Lloyd –
S.L.Lloyd@qmw.ac.uk
Abstract
This addendum contains additional information with respect to the bid from the UK Particle Physics Community to PPARC for resources to develop a Grid for Particle Physics research - GridPP. It should be read in conjunction with the proposal and contains responses in the following areas:
Table of Contents
1. Resource Allocation
*1.1 Introduction
*1.2 Funding Scenarios
*1.2.1 £20M Budget
*1.2.2 £17M Budget
*1.2.3 £15M Budget
*1.3 Experiment Objectives
*2. GridPP Financial Comparison with other Grid Projects
*2.1 US
*2.1.1 CMS
*2.1.2 ATLAS
*2.2 France
*2.3 Germany
*2.4 Italy
*3. International Grid Collaborations
*3.1 General/Summary
*3.2 GriPhyN
*3.3 PPDG
*3.4 iVDGL
*3.5 InterGrid Co-ordination Group
*3.5.1 Structure
*3.5.2 Aims
*3.5.3 Open Source License
*3.6 Common Testbeds
*3.7 Future
*4. Grid Architecture
*4.1 GGF Grid Protocol Architecture Working Group
*4.1.1 Draft Charter
*4.1.2 Proposed Goals
*5. Links with Industry
*5.1 Introduction
*5.2 IBM-UK
*6. APPENDIX
*6.1 Deliverable Requirements
*The programme for the UK Grid for Particle Physics consists of a series of interlocking "components". Components 1 to 4 are considered to be essential to meet the requirements of the GridPP programme.
Table 1 shows the PPARC funds invested as a function of component to CERN and the UK.
Table 1: Top-level financial breakdown.
Component |
Description |
CERN contribution from PPARC (£M) |
Cost to PPARC (£M) |
Total PPARC (£M) |
1 |
Foundation |
2.5 |
8.5 |
8.5 |
2 |
Production |
1.0 |
4.1 |
12.7 |
3 |
Middleware |
2.1 |
4.4 |
17.0 |
4 |
Exploitation |
1.5 |
4.0 |
21.0 |
5 |
Value-added exploitation |
2.9 |
4.9 |
25.9 |
Integrated over the first 4 components, the amount required for funding CERN activities is £7M.
The financial breakdown into various categories is given in Table 2 as a function of year.
Table 2: Financial breakdown by year for components 1-4.
Category |
to Apr-02 |
Costs to Apr-03 |
to Apr-04 |
Total (£M) |
Tier Centres Staff |
0.30 |
0.81 |
1.35 |
2.46 |
Capital |
1.06 |
0.87 |
1.03 |
2.96 |
Work Groups Staff |
1.24 |
3.38 |
3.21 |
7.84 |
Capital |
0.05 |
0.10 |
0.10 |
0.24 |
CERN Component Staff |
0.80 |
2.63 |
2.23 |
5.67 |
Capital |
0.38 |
0.47 |
0.56 |
1.42 |
Managers |
0.08 |
0.16 |
0.16 |
0.41 |
Totals |
3.91 |
8.43 |
8.64 |
20.98 |
Clearly the total of £3.91M to Apr-02 is not within the range of £2-2.5M currently discussed so a certain amount of re-profiling will be necessary. However, since these numbers come from a bottom-up assessment of the effort needed to produce specified deliverables to a given time scale it is not sensible to perform a detailed re-profiling until the total project budget, and therefore the project scope, is defined. All the numbers in this document clarify the model presented in April. Once the funding profile is known the Project Management Board will revise the planning to optimise the project in the light of both financial and technical developments in the interim.
The Staff Year (and corresponding financial) breakdown into WorkGroups is given in Table 3 as a function of year, separated into the funding already committed by PPARC to the EU DataGrid (EUDG) and the PPARC funding for GridPP with the DataGrid component excluded (UKPP). The mapping between the GridPP WorkGroups and the EU DataGrid WorkPackages was provided in Table 11 of the Proposal. For some EUDG WorkPackages the mapping is not unique, in which case it is assigned to one WorkGroup for ease of identification. For example, WP8 maps onto WG-H and WG-I and the total of 21 FTE of EUDG effort is assigned to WorkGroup H.
Table 3: Staff Years by WorkGroup as a function of year for components 1-4.
WorkGroup |
Year-1 |
Year-2 |
Year-3 |
Total |
||||
UKPP |
EUDG |
UKPP |
EUDG |
UKPP |
EUDG |
UKPP |
EUDG |
|
A |
1.12 |
0.50 |
2.15 |
0.25 |
2.13 |
0.75 |
5.40 |
1.50 |
B |
0.70 |
0.51 |
2.20 |
4.11 |
3.10 |
2.88 |
6.00 |
7.50 |
C |
0.00 |
1.50 |
1.80 |
0.00 |
2.40 |
0.90 |
4.20 |
2.40 |
D |
2.25 |
0.00 |
3.50 |
0.75 |
4.75 |
0.75 |
10.50 |
1.50 |
D |
0.00 |
1.60 |
0.00 |
1.45 |
0.00 |
1.45 |
0.00 |
4.50 |
E |
1.10 |
0.00 |
2.66 |
0.00 |
2.70 |
0.00 |
6.46 |
0.00 |
F |
1.70 |
1.13 |
3.07 |
2.27 |
2.73 |
2.60 |
7.50 |
6.00 |
G |
3.50 |
0.00 |
14.00 |
0.00 |
20.00 |
0.00 |
37.50 |
0.00 |
H |
1.50 |
4.93 |
5.25 |
8.03 |
5.75 |
8.03 |
12.50 |
21.00 |
I |
4.50 |
0.00 |
21.18 |
0.00 |
20.50 |
0.00 |
46.18 |
0.00 |
J |
2.00 |
0.00 |
5.00 |
0.00 |
3.00 |
0.00 |
10.00 |
0.00 |
Totals (SY) |
18.37 |
10.18 |
60.81 |
16.86 |
67.06 |
17.36 |
146.24 |
44.40 |
28.55 |
77.67 |
84.43 |
190.64 |
|||||
Totals (£M) |
0.99 |
0.55 |
3.28 |
0.91 |
3.62 |
0.94 |
7.89 |
2.40 |
1.54 |
4.19 |
4.56 |
10.29 |
A further breakdown according to the deliverables described in the Proposal is given in Table 5 of the Appendix, incorporating CERN-based activities and external funding from the EU.
We have been requested to plan for a programme in the range of £15-20M. This is at the point where elements of the Component 3 (Middleware) and Component 4 (Exploitation) need to be considered.
Component 3 enables the UK and CERN to make suitable contributions towards generating the middleware required for full Grid functionality, and in the process create generic software which will be used by other disciplines and in wider contexts. Component 4 is focused on providing the applications to deliver the science using the Grid infrastructure.
Components 1 to 4 must be funded, as an absolute minimum, in order for the GridPP programme to be realised
Figure 1: Breakdown of GridPP Resources for components 1-4 (£21M). The WorkGroups are: A Workload Management; B Information Services and Data Management; C Monitoring Services; D Fabric Management and Mass Storage; E Security Development; F Network Development; G Prototype Grid; H Software Support; I Experimental Objectives; J Dissemination; K CERN. The EU DataGrid components are denoted by an asterisk.
In the budget scenarios discussed below, we make the assumption that the total CERN component scales with the total but that the PPARC commitment to the EU DataGrid cannot be reduced. Furthermore, we apply a threshold below which we do not feel it sensible to arbitrarily scale elements of the project. Reductions beyond this threshold are achieved by reducing the scope of the project.
Figure 2: Breakdown of GridPP Resources for a £20M Budget. The EU DataGrid components (already funded) are displaced. The reductions and scaling factors discussed in the text for the non-EU DataGrid components are given.
We have examined the impact of a £17M program and conclude that it is not possible to simply scale back the existing Proposal. The solution we would propose is to reduce, pro-rata, the CERN component to £6M and the UK capital spend to £2.45M giving a total saving of £1.85M. We would not scale back the commitments to the EU DataGrid, but the other elements of the programme except for the experiment objectives (WG-I) would be funded at an average level of 90%, saving a further £0.54M. This would require input from the annual review mechanisms that are envisaged within the programme. Once a level of funding is approved, the detailed work packages will be put out to tender by the GridPP project management and all groups will be invited to bid for this work. Bids will be fully peer reviewed and will be selected on the basis of competence, commitment and current involvement of the groups. To achieve the total programme cost of £17.3M, a further savings of £1.3M is made by cutting in half Component 4 (Experiment Objectives), which would be reduced from £2.5M to £1.2M. In practice, this is a false economy since this work would have to be funded by the experiments from other elements of the core programme.
Figure 3: Breakdown of GridPP Resources for a £17M Budget. The EU DataGrid components (already funded) are displaced. The reductions and scaling factors discussed in the text for the non-EU DataGrid components are given.
Within the component model, it is impossible to achieve the programme described in the Proposal with £15M. The CERN component would have to be scaled back to £5M, saving an additional £1M. The other £1.2M saving would have to come from eliminating the experimental objectives (Component 4) completely from the scope of this project. The overall balance of the GridPP project would be radically distorted. We stress that the concurrent development of the applications is essential to provide feedback during the development of the Grid.
Figure 4: Breakdown of GridPP Resources for a £15M Budget. The EU DataGrid components (already funded) are displaced. The reductions and scaling factors discussed in the text for the non-EU DataGrid components are given.
The computing requirements of the LHC were described in the Proposal: development of the storage, management, simulation, reconstruction, distribution and analysis of the data of the four LHC experiments (ALICE, ATLAS, CMS and LHCb) constitutes an unprecedented challenge to the High Energy Physics (HEP) and Information Technology (IT) communities. The design of the software, both experimental and generic, to exploit Grid computing needs to start now in order to evaluate fully its performance and assess whether it will meet future demands. In parallel to addressing these demands, it is also crucial that the experiments generate large amounts of Monte Carlo simulated data in order to understand and optimise the design of the detectors to maximise the physics return in 2006. Both these needs call for substantial investment in computing infrastructure, both in the area of computing hardware and software development.
Less emphasis was placed on the integration of the current US experiments in the Proposal: the concurrent development of applications driven by real data, with immediate requirements of Grid technologies is also important. Examples of current developments of this aspect of the programme are described below.
The D0 experiment has developed a sophisticated data-handling tool known as SAM. The UK is implementing job submission and resource management aspects into SAM; this involves incorporating a Grid enabled submission mechanism. The project is being carried out in the context of the PPDG project and in close collaboration with the Condor team, based at Wisconsin University. Other UK personnel on D0 are heavily involved with Grid interfaces to mass storage elements within PPDG. Contacts are already developing between this work and the UK led DataGrid work package in this area.
In the UK, BaBar will develop a Tier-A processing centre that will meet a significant fraction of the whole collaboration’s computing needs. Working directly with PPDG, it is intended to implement a version of the storage request broker (developed at the San Diego Supercomputer centre) that is Grid enabled at this centre. In addition, a UK distributed Monte Carlo production system is currently being commissioned based on 9 PC farms. It is planned that this production system will be integrated into the Grid environment. BaBar plan to use the PPDG/Globus tools for bulk data transfer and the UK currently co-ordinates the BaBar developments in this area.
CDF’s computing model was developed based on their Run-I experience and the assumption that network connectivity between Europe and the States would continue to be inadequate. The model was developed very early in the CDF upgrade plans and, as a consequence, a major investment in their software environment, outside of a distributed framework, was made. Even so, the UK is driving forward the CDF Grid effort in collaboration with US institutes. Early work has included making the CDF metadata available via LDAP so that each collaborating site can publish a list of available resources and datasets. A feasibility study has also been undertaken which shows that the existing Disk Inventory Manager can be made Grid aware in a simple way. Support through GridPP for the CDF developments would allow these efforts to meet their full potential and spawn possible future funding in the US for CDF Grid effort.
The UK is playing a leading role in developing the middleware of the DataGrid. It is equally important that the UK take advantage of their prominent positions with the experimental collaborations in order to ensure that GridPP is taking a leading role in stress testing and utilising the developing tools. Not only should the UK build on its leadership within the LHC experiments, in the context of the DataGrid project, but should also exploit its prominent position within the US collider experiments in the area of Grid computing. This would enable important links to be forged between the EU DataGrid project, through GridPP, and the GriPhyN and PPDG projects in the States; thus fostering a closer co-operation.
It is difficult to compare the Grid projects of other countries because of different policies on funding of staff and different profiles of spend. The following information has been determined from a meeting of the HEP-CCC in Bologna (June 15), where GridPP was represented, and focuses on the investment in Tier-1 Regional Centres and associated Tier-2 Centres in various countries. These show the UK plans as being similar to other countries with the most e-Science aware one, Italy, being significantly more aggressive.
US CMSThe US plans a Tier-1 and 5 Tier-2 centres for CMS. A prototype will be built during 2000-04, with full deployment during 2005-7 (30:30:40 funding profile)
Staff estimates for the Tier-1 centre are 14 FTE by 2003, reaching 35 FTE in 2007.
Total Costs for Tier-1 and Tier-2 for 2001-2004 are $M(3.5, 4, 5.5, 8.5).
The integrated costs up to 2006 are $54.7M for hardware and software support for CMS. In addition, GriPhyN and PPDG (described in section 3) will also contribute to US LHC work. The currently approved budgets are $12.5M for GriPhyN1, $25M for GriPhyN2.
ATLAS plans a similar structure of Tier-1 and 5 Tier-2 centres for the same cost as CMS. We did not have access to detailed costs but estimates of the size is the same as CMS for CPU, less for disk and more for tape.
France plans a Tier-1 RC for all 4 LHC experiments at CC-IN2P3 in Lyon and an LHC prototype starting now. The size is similar to a prototype Tier-1 defined in the CERN Hoffmann report. Estimates but no details of manpower have been determined for IN2P3.
2M€/year for a National Grid has been assigned now by ministerial action. CC-IN2P3 are also working on a BaBar Tier-A centre and Grid work with BaBar. The Tier-1/A work will be a major part of CC-IN2P3 work in future years. Their annual budget is 8M€ and 40FTE.
Karlsruhe run a large computer centre for a variety of disciplines (50FTE and 11M€ annual budget). The Tier-1 RC for Germany will be sited there. It will make use of existing infrastructure with 8 additional staff and 12M€ for hardware assigned up to 2006. This is for Tier-1 only and provides no experimental support.
Italy plans a Tier-1 RC and a prototype starting now in CNAF, Bologna. Hardware plans are more aggressive than UK. 15.9M€ is allocated during 2001-3 for Tier-1 hardware and outsourced staff. Additional staff will be funded for Tier-1 support (7; 21; 25 FTE for each year). They also plan 10 Tier-2 centres at 1M€/year. An initial 24M€ budget has already been allocated.
GriPhyN is the major Grid technology development project in the US, which embodies Particle Physics applications as well as other fields.
We have spoken in various contexts with the GriPhyN management and to individuals working in specific areas and have identified the following areas:
PPDG focuses more strongly on HEP applications, and particularly the near term requirements of CDF, D0 and BaBar. The respective GridPP groups are naturally an integral part of this work. All experiments have significant data handling requirements already, and will continue to do so for several years.
The iVDGL is an essential part of the US HEP Grid programme, which we support and expect to be funded. The iVDGL seeks to link similar Grid domains on a global scale. The DataGrid and GridPP see this as the correct and essential way forward. So far:
This will:
"International HENP Grid Coordination & Joint Development Framework"
Strategic co-ordination of Particle Physics Grid projects has been the subject of meetings between representatives from EU-DataGrid, GriPhyN, PPDG and Asian activities. So far these have taken place in Amsterdam (March 2001) and Rome (June 2001). The UK has been represented by GridPP and by PPARC. It is foreseen that in future PPARC will be represented by the Director e-Science. The group also has amongst its members the Principal Investigators of the Globus and Condor projects. The next meeting is foreseen to be in October.
After much discussion it has been agreed to operate a 2-tier structure. The first part is a Management Board responsible for overall co-ordination. The second part is a Technical Board charged with ensuring that all technical issues are covered either by launching task forces of domain experts, by ensuring joint activities of existing projects or by requesting new (joint) projects (the latter passed to the Management Board for approval and initiation). The exact terms of reference of these two bodies are currently being developed.
An additional tier of common projects management has also been discussed, but deemed to be of limited benefit, since such bodies would have no direct resources but would in any case rely on the funding available through core participants.
The existing body is to become the Management Board and nominations are currently sought for the Technical Board (6 US, 6 Europe, 2 Asia) in order that it may start its work before the end of the summer. GridPP has proposed one member to sit on this technical board.
The overall aim of the co-ordination is primarily to promote common activities with a view to ensuring seamless interfaces for applications in the various countries and to avoid (or at least minimising) duplication. To date it has been decided to promote joint testbeds, grid architecture co-ordination and to work for a common Open Source license structure. The DataTAG project and its association with iVDGL is in part a consequence of this co-ordination activity.
Work has also progressed on software licensing issues with an outline for a Consortium for Open Grid Software (COGS) having been discussed. Associated Open Source license text is already available and has been distributed to participating institutes of GridPP for comment and feedback. It is foreseen that developments of the EU DataGrid will be made available under this license.
As part of the first testbed release of the EU DataGrid project it is planned to have the direct participation of up to 4 sites in the US in the testbed and is anticipated to include at least the prototype Tier-1 centres. The GridPP prototype activities will this be fully linked with this.
The general policy, as GridPP gets underway, will be to identify the Grid components required to meet the GridPP goals. We will then actively seek to coordinate closely with the US grid projects in order to benefit from existing R&D and avoid unnecessary duplication. We expect this approach to enable further mutually beneficial collaboration.
Note this policy will also naturally include other UK resources outside of GridPP such as the National and Regional e-Science Centre infrastructure.
To a great extent this stated policy will be formalised through participation in the InterGrid Co-ordination Group, whose job will be to ensure proper collaboration takes place wherever appropriate.
Grid ArchitectureOur overall aim is to make sure that the Grid is a reality for PP in the UK: we feel that the best way to achieve this is by making individual contributions to architectures via our links to the Global Grid Forum (GGF), EU DataGrid, PPDG and GriPhyN. GridPP does not plan to have its own architecture body.
We support the GGF in its attempts to promote best practices and to establish standards and will take particular note of the activities of the recently formed architecture group within the GGF. We expect to see standards emerge over the next few years and we are especially concerned that architectures should be capable of evolution.
We have a strong link to the Architecture Task Force (ATF) of the EU DataGrid, with three UK members on that body, one of whom is currently the chair.
We also have direct links to the National e-Science Centre and five of the eight Regional e-Science Centres, which will provide a forum for architecture development in the future. The National Centre Director and the Chief Software Architect who supports the development of the research strategy will have an important role in ensuring the evolutionary capabilities of the architecture.
The GGF Grid Protocol Architecture Working Group draft charter is noted below. Various members of GridPP will be attending the GGF meeting in Washington and contributing to discussions in the Grid Protocol Architecture Working Group. The GridPP project leader is a member of this Working Group and will provide a link to ensure that the architecture adopted by GridPP is fully consistent with that developed in the Grid Forum.
GGF Grid Protocol Architecture Working Group Draft CharterThe role of the Grid Protocol Architecture Working Group is to provide a conceptual framework for discussing the interrelationships, completeness, and minimality of the protocol approach to Grid services that is coming out of GF.
Examine and document the relationship of the GPA architecture with respect to other approaches such as CORBA, peer-to-peer, etc.
Links with Industry IntroductionThe PP community intends to build upon its existing links with industry. We intend to pursue these links at several levels. We will develop high level relationships through our Dissemination Board and grass roots contacts through our Regional Centres and Institutes. Industrial partners are already working with PP groups in JREI and JIF projects, in particular in the provision of compute-engines and mass-storage facilities. These facilities are part of the core infrastructure for the national GridPP, and the industrial partners have joined with us to gain from our experiences. The current investment in hardware and associated industrial partnership is shown in Table 4.
Table 4: External Funds (additional to PPARC Grants and central facilities) providing computing equipment and the basis for Industrial Partnership
Collaboration |
Cost (£M) |
Industrial Partner |
Funding |
BaBar (Birmingham, Bristol, Brunel, Edinburgh, Imperial, Liverpool, Manchester, QMUL, RAL, RHUL) |
0.8 1.0 |
Sun |
JREI JIF |
MAP (Liverpool ) |
0.3 |
ITS |
JREI |
ScotGrid (Edinburgh, Glasgow) |
0.8 |
IBM |
JREI |
D0 (Lancaster) |
0.4 0.1 |
WS-UK |
JREI Univ. |
Dark Matter (Sheffield) |
0.03 |
Tbd |
JIF |
CDF/Minos (Glasgow, Liverpool, Oxford, UCL) |
1.7 |
IBM |
JIF |
CMS (Imperial) |
0.15 |
Tbd |
JREI |
ALICE (Birmingham) |
0.15 |
Tbd |
JREI |
Total Cost |
5.4 |
Our aims with respect to industrial partnership are to: help us develop the Grid by supplying hardware and software; use the Grid for themselves for collaborative engineering, massive simulations and federating their own worldwide databases; sell or develop the Grid for other early adopters. This links to UK developments via the National e-Science Centre: we are keen to ensure that each of the GridPP Regional Centres benefits from a co-ordinated approach to industrial partnership, a task which will be taken up by the Dissemination Board.
As part of its growing commitment to the Grid, IBM has dedicated resources through various collaborations with Institutes throughout the UK. The process of identifying appropriate Grid tools from IBM’s existing strategic development areas in e-Utilities, e-business (project eLiza), e-markets and Intelligent Infrastructure with respect to the Grid architecture has started. Figure 5 illustrates this relation: there is a clear mapping of the Collective, Resource, Connectivity and Fabric layers to the Middleware WorkGroups in GridPP. Various Institutes are currently identifying areas of common interest.
Figure 5: Layers within the Grid mapping to IBM technologies. Each one of the layers will have its own Application Programming Interfaces (APIs) and Software Development Kits (SDKs).
An early example of a collaborative project with IBM-UK is the development of a Producer/Consumer model for the Grid Monitoring Architecture (GMA), interfaced to an SQL service. This relational model enables both live and archived monitoring data to be efficiently queried with a common interface. The GGF performance architecture does not currently specify the protocol between the Producer and Consumer: the relational model has been developed consistent with the GMA and was recently demonstrated at the Oxford EU DataGrid Meeting.
We envisage a close relationship in the development of the Fabric, Connectivity and Resource layers with vendors for the UK Tier centres. We also envisage relationships with database providers such as Objectivity, Oracle, MySQL and PostgreSQL from the Application and Collective layers. The PP community will take the best advantage of the opportunities of industrial liaison to enhance GridPP, and in turn will pro-actively seek to aid the transfer of technologies, rather than being a passive application driver.
Table 5 is a compilation of the deliverables from all the WorkGroups, ordered by component, and updated with respect to Table 3 of the Proposal to include individual staff requirements in units of Staff Years (SY) matched to each of the deliverables. The deliverable definitions are considered in more detail in Section 14 of the Proposal.
Table 5: Top-level Deliverables including Staff Year requirements.
Component |
WG |
Name |
Funding |
SY |
1 |
A |
Installation and test job submission via scheduler |
UKDG |
0.50 |
1 |
A |
Planning and Management |
UKPP |
1.50 |
1 |
A |
Develop JCL/JDL |
UKPP |
0.50 |
1 |
B |
Develop Project Plan, Coord. + Manage |
EU |
1.50 |
1 |
B |
Schema Repository |
UKDG |
1.50 |
1 |
B |
Releases A |
UKDG |
0.50 |
1 |
C |
Release for Testbed-1 |
EU |
1.50 |
1 |
C |
Release for Testbed-2 |
EU |
1.50 |
1 |
C |
Release for Testbed-3 |
EU |
1.20 |
1 |
C |
Evaluation Report |
EU |
0.30 |
1 |
C |
Evaluation Report |
UKDG |
0.90 |
1 |
D |
Evaluation and API Design |
EU |
0.50 |
1 |
D |
Prototype API |
EU |
0.50 |
1 |
D |
Further Refinement and testing of API |
EU |
1.00 |
1 |
D |
COTS systems development B |
UKDG |
1.50 |
1 |
D |
Definition of Metadata |
UKDG |
0.30 |
1 |
D |
Prototype Metadata |
UKDG |
0.30 |
1 |
D |
Metadata refinement and testing |
UKDG |
0.70 |
1 |
D |
Tape Exchange Prototype Version |
UKDG |
0.75 |
1 |
D |
Develop project plan |
UKPP |
0.10 |
1 |
D |
Integration of existing fabric |
UKPP |
3.50 |
1 |
D |
Fabric benchmarking/evaluation |
UKPP |
1.00 |
1 |
D |
User Portals |
UKPP |
0.50 |
1 |
D |
Fabric demonstrator(s) |
UKPP |
1.00 |
1 |
E |
Gather requirements |
UKPP |
0.20 |
1 |
E |
Survey and track technology |
UKPP |
0.50 |
1 |
E |
Design, implement and test |
UKPP |
1.60 |
1 |
E |
Integrate with other WG/Grids |
UKPP |
0.66 |
1 |
E |
Management of WG |
UKPP |
0.25 |
1 |
E |
DataGrid Security |
UKPP |
0.50 |
1 |
F |
Net-1-A |
UKDG |
1.00 |
1 |
F |
Net-2-A |
UKDG |
0.50 |
1 |
F |
Net-4-A |
UKDG |
1.50 |
1 |
F |
Net-1-B |
UKPP |
2.00 |
1 |
F |
Net-2-B |
UKPP |
1.00 |
1 |
F |
Net-4-B |
UKPP |
1.50 |
1 |
G |
Management |
EU |
3.00 |
1 |
G |
GRID IS |
UKPP |
4.50 |
1 |
G |
Network ops |
UKPP |
1.50 |
1 |
G |
Tier-1 centre ops |
UKPP |
17.50 |
1 |
H |
Deployment tools |
UKDG |
2.00 |
1 |
H |
Globus support |
UKDG |
2.00 |
1 |
H |
Testbed team |
UKDG |
1.50 |
1 |
H |
Management |
UKDG |
3.00 |
1 |
H |
Deployment tools |
UKPP |
2.50 |
1 |
H |
Globus support |
UKPP |
4.00 |
1 |
H |
Testbed team |
UKPP |
1.50 |
1 |
J |
Begin foundational package A |
UKPP |
2.00 |
1 |
K |
Support prototypes |
CERN |
3.00 |
1 |
K |
Extension of Castor for LHC capacity, performance |
CERN |
5.00 |
1 |
K |
Support prototypes |
CERN |
3.00 |
1 |
K |
Fabric network management, and resilience |
CERN |
2.00 |
1 |
K |
Support fabric prototypes |
CERN |
2.00 |
1 |
K |
High bandwidth WAN – file transfer/access performance |
CERN |
3.00 |
1 |
K |
WAN traffic instrumentation & monitoring |
CERN |
2.00 |
1 |
K |
Grid authentication – PKI |
CERN |
1.00 |
1 |
K |
Authorisation infrastructure for Grid applications – PMI |
CERN |
2.00 |
1 |
K |
Base technology for collaborative tools |
CERN |
2.00 |
1 |
K |
Support for grid prototypes |
CERN |
2.00 |
1 |
K |
Evaluation of emerging object relational technology |
CERN |
2.00 |
2 |
A |
Installation and test job submission via scheduler |
UKPP |
0.40 |
2 |
B |
Query Optimsation and Data Mining A |
UKPP |
0.90 |
2 |
C |
Technology Evaluation |
UKDG |
0.50 |
2 |
C |
Evaluation Report |
UKPP |
0.30 |
2 |
D |
Implementation of production metadata |
EU |
0.70 |
2 |
D |
Implementation of Production API |
UKDG |
1.00 |
2 |
E |
Production phase |
UKPP |
1.00 |
2 |
F |
Net-2-C |
UKDG |
0.50 |
2 |
F |
Net-3-A |
UKDG |
1.00 |
2 |
F |
Net-2-D |
UKPP |
1.00 |
2 |
F |
Net-2-G |
UKPP |
1.00 |
2 |
G |
Security operations |
UKPP |
3.00 |
2 |
G |
GRID IS |
UKPP |
1.50 |
2 |
G |
Tier-1 centre ops |
UKPP |
9.50 |
2 |
H |
Upper middleware/application support |
UKDG |
5.50 |
2 |
H |
Upper middleware/application support |
UKPP |
4.50 |
2 |
J |
Focus and engage A |
UKPP |
2.00 |
2 |
K |
LAN performance |
CERN |
1.00 |
2 |
K |
High bandwidth firewall/defences |
CERN |
2.00 |
3 |
A |
Further testing and refinement |
UKDG |
1.00 |
3 |
A |
Modify SAM |
UKPP |
1.00 |
3 |
A |
Profiling HEP jobs and scheduler optimisation |
UKPP |
1.50 |
3 |
A |
Super scheduler development |
UKPP |
0.50 |
3 |
B |
Directory Services |
EU |
3.00 |
3 |
B |
Distributed SQL Development |
UKDG |
4.00 |
3 |
B |
Data Replication |
UKDG |
1.50 |
3 |
B |
Query Optimsation and Data Mining B |
UKPP |
0.60 |
3 |
B |
Releases B |
UKPP |
2.00 |
3 |
B |
Liason |
UKPP |
2.50 |
3 |
C |
Architecture & Design |
UKDG |
0.50 |
3 |
C |
Technology Evaluation |
UKDG |
0.50 |
3 |
C |
Release for Testbed-2 |
UKPP |
1.80 |
3 |
C |
Release for Testbed-3 |
UKPP |
2.10 |
3 |
D |
Tape Exchange Production Version |
EU |
0.30 |
3 |
D |
Tape Exchange evaluation & design |
UKDG |
0.50 |
3 |
D |
Design Refinement |
UKDG |
0.50 |
3 |
D |
Tape Exchange Production Version |
UKDG |
0.45 |
3 |
D |
Fabric Management Model |
UKPP |
0.10 |
3 |
D |
Establish ICT-industry leader partnerships |
UKPP |
0.10 |
3 |
D |
COTS systems development A |
UKPP |
1.50 |
3 |
D |
Proprietary systems development |
UKPP |
1.00 |
3 |
D |
FM information dissemination |
UKPP |
0.20 |
3 |
D |
Evaluation Report |
UKPP |
1.50 |
3 |
E |
Architecture |
UKPP |
0.25 |
3 |
E |
Security development |
UKPP |
0.75 |
3 |
E |
DataGrid Security development |
UKPP |
0.75 |
3 |
F |
Net-2-E |
UKDG |
1.00 |
3 |
F |
Net-3-B |
UKDG |
0.50 |
3 |
F |
Net-2-F |
UKPP |
1.00 |
3 |
H |
Globus development |
UKDG |
1.50 |
3 |
H |
S/w development support |
UKDG |
1.50 |
3 |
H |
Upper middleware/application support |
UKDG |
4.00 |
3 |
J |
Begin production phase A |
UKPP |
3.00 |
3 |
J |
QCDGrid – full Grid access of lattice datasets |
UKPP |
3.00 |
3 |
K |
Scalable fabric error and performance monitoring system |
CERN |
1.00 |
3 |
K |
Automated, scalable installation system |
CERN |
2.00 |
3 |
K |
Automated software maintenance system |
CERN |
1.00 |
3 |
K |
Scalable, automated (re-)configuration system |
CERN |
1.00 |
3 |
K |
Automated, self-diagnosing and repair system |
CERN |
2.00 |
3 |
K |
Implement grid-standard APIs, meta-data formats |
CERN |
2.00 |
3 |
K |
Data replication and synchronisation |
CERN |
3.00 |
3 |
K |
Performance and monitoring of wide area data transfer |
CERN |
3.00 |
3 |
K |
Integration of LAN and Grid-level monitoring |
CERN |
1.00 |
3 |
K |
Adaptation of databases to Grid replication and caching |
CERN |
5.00 |
3 |
K |
Preparation of training courses, material |
CERN |
4.00 |
3 |
K |
Adaptation of application – science A |
CERN |
3.00 |
3 |
K |
Adaptation of application – science B |
CERN |
3.00 |
4 |
I |
ATLAS |
UKPP |
8.00 |
4 |
I |
CMS |
UKPP |
9.50 |
4 |
I |
LHCb |
UKPP |
9.00 |
4 |
I |
ALICE |
UKPP |
2.50 |
4 |
I |
BaBar |
UKPP |
6.00 |
4 |
I |
UKDMC |
UKPP |
3.00 |
4 |
I |
H1 |
UKPP |
1.50 |
4 |
I |
CDF |
UKPP |
4.68 |
4 |
I |
D0 |
UKPP |
2.00 |
4 |
I |
ZEUS |
UKPP |
0.00 |
4 |
K |
Provision of basic physics environment for prototypes |
CERN |
2.00 |
4 |
K |
Support of grid testbeds |
CERN |
5.00 |
4 |
K |
Adaptation of physics core software to the grid environment |
CERN |
6.00 |
4 |
K |
Exploitation of the grid environment by physics applications |
CERN |
6.00 |
4 |
K |
Support for testbeds |
CERN |
3.00 |