Infrastructure and Basic Grid Services

29
Infrastructure and Basic Grid Services Working Group Kick-Off Meeting SWITCH, 2008 January 15

description

Infrastructure and Basic Grid Services. Working Group Kick-Off Meeting SWITCH, 2008 January 15. Agenda. 09:45 - 10:15: Charter of the WG, relation to EB and other WGs - PowerPoint PPT Presentation

Transcript of Infrastructure and Basic Grid Services

Page 1: Infrastructure and Basic Grid Services

Infrastructure and Basic Grid Services

Working Group Kick-Off MeetingSWITCH, 2008 January 15

Page 2: Infrastructure and Basic Grid Services

2008 January 15 2Infrastructure and Basic Grid Services WG

Agenda

09:45 - 10:15: Charter of the WG, relation to EB and other WGs

10:15 - 11: 45 : From every institution a short summary what they currently have installed and what they are planning to do in 2008.

11:45 - 12:30: Exchange of middleware experiences: Every participant, who has first hand experiences with a certain grid middleware gives a short presentation with the strengths and weaknesses of this particular middleware from his point of view.

12:30 - 13:30 Lunch

13:30 - 14:30: Discussion, which middlewares should be supported and how

14:30 - 16:00: Next steps: Identification of sub-projects and assignment to groups of 3-4 persons

Page 3: Infrastructure and Basic Grid Services

2008 January 15 3Infrastructure and Basic Grid Services WG

Charter – Focus/Purpose

One of the mission statements of SwiNG is the establishment of a sustained Swiss Grid Infrastructure. There are two basic components necessary to do so:

1. a set of shared hardware resources (computing, data and instruments) and

2. a set of Grid services linking the resources and making them accessible for use.

The resources available to the Swiss Grid Infrastructure are provided by the individual member institutions and their sub-units. Concerning the first point, the manner by which the resources are shared has to be established and adapted to the constraints of each individual partner. In addition, mechanisms have to be set up by which sufficient incentive is created to share and keep sharing the resources, as well as to attract new shared resources.

Concerning the second point, the basic Grid middleware services that are independent of the applications running on the Grid have to be set up and operated by the SwiNG member institutions. The services have also to be defined, categorized, monitored and maintained, and the exact mechanisms how to do so defined.

Page 4: Infrastructure and Basic Grid Services

2008 January 15 4Infrastructure and Basic Grid Services WG

Scope

This activity builds on previous Swiss effort [R1-2]. In the scope of SwiNG we solicit the active participation of all members that have any kind of shareable resource. The initial work of the WG will define the detailed mechanisms of the resource sharing. This includes drafting the necessary service level agreements for the participating partners, the detailed mechanisms of accounting, access, monitoring, etc. Processes have to be defined and set up by which middleware is selected, deployed and operated.

Strategic decisions concerning the infrastructure and basic Grid services are taken on the level of the Association, i.e. the Executive Board and the Assembly. The WG is responsible to prepare the necessary material for such decisions.

The ISBG-WG may re-scope itself at a later time into smaller subsets later, once the bootstrapping of the infrastructure is done and the processes are understood.

The scientific requirements will be defined by other WGs but collected by us.

Page 5: Infrastructure and Basic Grid Services

2008 January 15 5Infrastructure and Basic Grid Services WG

Goals

There are many examples of existing Grids and other countries’ national infrastructures that can be taken as basis for this WG. We are in the comfortable position to have a lot of experience to draft from ([R3-R7] and more), picking the best concepts for the Swiss academic landscape, which has its own peculiarities.

The WG is expected to define the processes and procedures to set up and operate the initial Swiss Grid Infrastructure, and actually putting them to practice. A very important aspect of the WG is to define the tasks that need dedicated funding to take place.

Page 6: Infrastructure and Basic Grid Services

2008 January 15 6Infrastructure and Basic Grid Services WG

Inventory

From every member group a short summary what they currently have installed and what they are planning to do in 2008.

• What is there in terms of resources (compute, data, instruments)

• Current projects

• Interests

• Goals for 2008

• Experience, lessons learned from previous projects

Page 7: Infrastructure and Basic Grid Services

2008 January 15 7Infrastructure and Basic Grid Services WG

SWITCH

• What is there in terms of resources (compute, data, instruments)– Grid Team is a subset of the Middleware team; 5 people (not all full-time)– gLite-based installation running central services

WMS, File Catalogs, VOMS, Info systems

• Current projects– EGEE-II activities – shibboleth interoperability in JRA1– Middleware security focus

• Goals for 2008 and beyond– EGEE-III starting this spring– Core focus of SWITCH in general is on the national scope, hence very active

in SwiNG, importance of national scope will grow– Provide central services

PKI AAI / SLCS Non-browser based applications for Shibboleth

– NO computing resources– AAA Grid Projects – define projects, attract funding

Page 8: Infrastructure and Basic Grid Services

2008 January 15 8Infrastructure and Basic Grid Services WG

PSI

• What is there in terms of resources (compute, data, instruments)– Using gLite, previously also AliEn

• Current projects– LCG Grid, Science Driven– Data analysis is extremely important (several PB/year to analyze)

System for data sharing Securing data access

• Goals for 2008– Detector is giving data this year, ramping up– Entering production this year, supply a stable production Grid– Building an LCG Tier3 at PSI, 120CPUs/70TB

• Interests– Provide access to local resources

• Experience, lessons learned from previous projects– Small dev team makes for flexible testing– Selection of supported requirements is important

There may be only few questions concerning science requirements (cpu, data, instrument access)

– gLite can cover data access aspects, but support is very expensive – don’t know of an alternative at this scale

Page 9: Infrastructure and Basic Grid Services

2008 January 15 9Infrastructure and Basic Grid Services WG

EPFL – DIT

• What is there in terms of resources (compute, data, instruments)– Desktop Grid with 300 CPUs multi-platform (45% linux, 50% windows,

rest MacOSX)– Condor-based, only available at nights and week-ends

• Current projects– Condor Desktop grid

• Interests– Providing resources to the users inside and outside

Users are physics and chemistry today

• Goals for 2008– Add more compute nodes to the Condor desktop grid (there are 3000

unused nodes yet)– Improve NorduGrid interface

• Experience, lessons learned from previous projects

Page 10: Infrastructure and Basic Grid Services

2008 January 15 10Infrastructure and Basic Grid Services WG

Uni BE - IT

• What is there in terms of resources (compute, data, instruments)– Ubelix cluster (500cores)– Several science users– NorduGrid is installed– gLite has been tried but is not available

• Current projects– Parallel filesystem: Lustre

• Interests– Providing standardized access

• Goals for 2008– Enlarge cluster with more hardware– Access to cluster should be standardized Swiss-wide

• Experience, lessons learned from previous projects

Page 11: Infrastructure and Basic Grid Services

2008 January 15 11Infrastructure and Basic Grid Services WG

FGCZ

• What is there in terms of resources (compute, data, instruments)– There are a lot of instruments (mass spectrometry, etc)– Cluster that is used is at the IT– Several TB of data

• Current projects– Data centric– B-Fabric infrastructure

Based on open source projects to do reliable data mgmt Store data and annotate data (metadata) Share data – security, access

– Avaki (sybase) However this is commercial now

• Interests– Data management, data sharing issues, not only CPU cycles– Build basic services allowing to provide secure access to distributed data– Promote B-Fabric

• Goals for 2008– Extension of B-Fabric wrt workflows, new instruments, etc– Working with Unilever R&D and CSI-D (they use B-Fabric)– Promote B-Fabric, establish a community, extend opensource subsystems (e.g Cocoon). Allows

generating new interfaces.• Experience, lessons learned from previous projects

– Avaki– Grid Transactions (GGF)

Page 12: Infrastructure and Basic Grid Services

2008 January 15 12Infrastructure and Basic Grid Services WG

UZH-OCI Comp chem and Grid computing

• What is there in terms of resources (compute, data, instruments) (all linux)– Dedicated cluster for Grid projects 2 frontend nodes 8 compute nodes (dual core) [TEST]– Special purpose servers– Production-style cluster > 100 nodes for comp.chem, not shared [PRODUCTION]– Dedicated infrastructure for SwissRe project– Globus4, Nordugrid, Unicore, and a lot more..– Websphare, DataSynapse

• Current projects– SwissRe collaboration on financial loss prediction (insurance)– SwiNG– PRAGMA– SUK project dealing with NorduGrid and AAI– SEPAC– Chemomentum (grid workflow based infrastructure) FP6 EU project– And a few more related projects like GEMSTONE, all related to comp.chemistry, workflows, web services, etc

• Interests– Provide infrastructure to adapt application code to Grid infrastructures– Provide a real production-style Grid infrastructure to user groups (usable and stable), bridge the gap between

application and infrastructure• Goals for 2008

– Double nodes for dedicated grid cluster– Continue SwiNG involvement and all other projects– Attract/secure more 3rd party funding

• Experience, lessons learned from previous projects– Funding drives everything– Data will become more important (complexity for us, not size)

Page 13: Infrastructure and Basic Grid Services

2008 January 15 13Infrastructure and Basic Grid Services WG

UZH-IT

• What is there in terms of resources (compute, data, instruments)– Hardware: 384dualCPU matterhorn cluster, attached experimentally to swiss

bio grid– United Devices grid (currently unused)

• Current projects– Small test-cluster (32x8 nodes) two different fast interconnects, possibly can

be used for MPI-grid testing + 10TB disk• Interests

– Serve the users– Make grid services accessible for non-experts (biologists, economists, ..)– Applications for UD– MPI on the grid

• Goals for 2008– Build up 2000CPU(or core) cluster for medicine faculty– Make a few blade centers available and accessible by grid users– Consolidation of data storage infrastructure, hierarchical storage managers

• Experience, lessons learned from previous projects– UD is administrator friendly but not coder friendly– 64bit opteron – seems still be problematic for Nordugrid and shibboleth

Page 14: Infrastructure and Basic Grid Services

2008 January 15 14Infrastructure and Basic Grid Services WG

UniGe – HEP

• What is there in terms of resources (compute, data, instruments)– 61 computers, 184cores . Fileservers, 3 service nodes. 75TB disk.

– SLC4 plus solaris. Using Torque and Maui, with Nordugrid

• Current projects– LCG-ATLAS

• Interests–Willingness to share is there

• Goals for 2008– Get the current system into production.

• Experience, lessons learned from previous projects

Page 15: Infrastructure and Basic Grid Services

2008 January 15 15Infrastructure and Basic Grid Services WG

UniBe – LHEP

• What is there in terms of resources (compute, data, instruments)– 34 worker nodes on SLC4, NorduGrid frontend, 10TB, and two more

fileservers with 22TB

– Rely on the IT services (Ubelix)

• Current projects– LCG-ATLAS

• Interests– Increase visibility our network, and the visibility of our work

– Access to Grid services that may be needed in the future, so that we don’t need to run them ourselves

– Access to more resources if possible, sharing own resources as possible (depends on effort as well)

• Goals for 2008– Have the Swiss computing infrastructure ready for data taking this year to

ATLAS

• Experience, lessons learned from previous projects

Page 16: Infrastructure and Basic Grid Services

2008 January 15 16Infrastructure and Basic Grid Services WG

HES-SO EIG

• What is there in terms of resources (compute, data, instruments)– 3 people working half-time. 300CPU windows/linux available 24/7

• Current projects– Development of XtremWebCH– Porting / developing applications on this platform (GAMESS, CONES,

Phylip, Neuroweb)

• Interests– Provide these resources to all applications (life science and medical

in particular)– Integration of XtremWebCH with other mw, interoperability

• Goals for 2008– Extensions to XtremWebCH– Bridging with gLite

• Experience, lessons learned from previous projects

Page 17: Infrastructure and Basic Grid Services

2008 January 15 17Infrastructure and Basic Grid Services WG

USI

• What is there in terms of resources (compute, data, instruments)– a few large servers – one small Apple cluster (16CPU)

• Current projects– EU-IP RESERVOIR (Virtualization, SLA) – JOpera grid workflow tool (http://www.jopera.org)

• Interests– Use available SwiNG resources– SwiNG for real-world requirements for research – Grid middleware research: Meta-middleware, Grid workflows – Integrate JOpera with more low-level Grid middleware (Condor already

supported) – Integrate JOpera with data management middleware

• Goals for 2008– Expand hardware resources at USI – Transfer research into "production" – Start SWITCH/AAA collaboration projects – Develop the JOpera Swiss user community

• Experience, lessons learned from previous projects

Page 18: Infrastructure and Basic Grid Services

2008 January 15 18Infrastructure and Basic Grid Services WG

CSCS

• What is there in terms of resources (compute, data, instruments)– Some service nodes and a few WNs (Biogrid + seed)

• Current projects– LCG / EGEE– PRACE– SEPAC

• Interests– Establish a national network to access the CSCS resources– Establishment of a high-throughput infrastructure in switzerland

• Goals for 2008– 25WN 2x2, 18TB, Suse– Successfully run the projects– Start the establishment of the network

• Experience, lessons learned from previous projects

Page 19: Infrastructure and Basic Grid Services

2008 January 15 19Infrastructure and Basic Grid Services WG

SIB - VitalIT

• What is there in terms of resources (compute, data, instruments)– 400CPU dedicated to life sciencies– 37TB storage, to be increased to 100TB by end of january– DB services (mysql, oracle), public DB repositories for genomics, ..– Dedicated systems with large memory– NodruGrid installed, gLite tried (constrained by OS)– LUSTRE

• Current projects– Swiss Bio Grid– EMBRACE

• Interests– DataGrid – storage services to be provided to our partners– Propagate the DB repository to other sites

• Goals for 2008– Accommodate high-throughput sequencing on the infrastructure – increase

storage by a large amount– Upgrade infrastructure further

• Experience, lessons learned from previous projects– Grid protocols are very firewall unfriendly

Page 20: Infrastructure and Basic Grid Services

2008 January 15 20Infrastructure and Basic Grid Services WG

Experience / Discussions

Current Middleware knowledge– Globus and globus-based

Globus 4 edg, gLite NorduGrid

– Condor– Unicore– Desktop Grids

BOINC United Devices XtremWebCH

Interoperability / Interoperation is deemed essential between various MW– Already working for some, like Condor/NorduGrid, gLite/NorduGrid

User Groups – definition is nontrivial– Scientists (focusing on publications)– Developers of services– Maintainers and supporters of resources

Applications– Predefined validated set of applications– User-compiled custom code– Combination of both

KISS principle – Keep It Simple and Stupid– For overall design of basic Grid services– File catalog, hierarchies– Authentication and authorization – either in the catalog, propagating a token or at the site that then needs to be synchronized.– Monitoring– Pilot Jobs and overlay grids

Page 21: Infrastructure and Basic Grid Services

2008 January 15 21Infrastructure and Basic Grid Services WG

Project Ideas

• Central Services:– Comparaison of resource brokers– Comparaison of IS– Monitoring, site functional tests– VO support

• Push vs Pull architectures• Site specific issues:

– Installation of different middlewares at the same site Compatibilities, problems

– Recommendations on LRMS– Use of AAI attributes for authorization

• Data Management Issues– Comparaison of different solutions – Identify user communities and their requirements

Page 22: Infrastructure and Basic Grid Services

2008 January 15 22Infrastructure and Basic Grid Services WG

File Management Project

Design the necessary components to provide distributed file access on the Grid.

• Store files

• Access files

Page 23: Infrastructure and Basic Grid Services

2008 January 15 23Infrastructure and Basic Grid Services WG

0. One Grid software

One selected middleware

To be installed everywhere

Choice needs to be made

Page 24: Infrastructure and Basic Grid Services

2008 January 15 24Infrastructure and Basic Grid Services WG

1. Infrastructure dependent projects

Every institution makes its choice

Minimal common elements

Independent infrastructures

Page 25: Infrastructure and Basic Grid Services

2008 January 15 25Infrastructure and Basic Grid Services WG

2. Centralization at the top

Sites are independent

Standardization effort at the top

Independent projects at the site

Common project at the top

Page 26: Infrastructure and Basic Grid Services

2008 January 15 26Infrastructure and Basic Grid Services WG

3. Common decentralized approach

Support of different stacks at the site

Standardization at the top as much as possible

Coordinated projects at site and top

Page 27: Infrastructure and Basic Grid Services

2008 January 15 27Infrastructure and Basic Grid Services WG

4. Common layer at the bottom

Common layer at every site

X may be very thin, not necessarily excluding G, N

In addition site specific / infrastructure specific components

Pull vs push architecture

Page 28: Infrastructure and Basic Grid Services

2008 January 15 28Infrastructure and Basic Grid Services WG

Data Management Project

• Project Idea:– Select one data management solution

– Install at a number of sites for supporting various Vos

– Could be used together with a grid middleware or independent of a given middleware

• Technical evaluation of DM solutions

• Gathering user requirements from scientific communities

• Implementation as a project between different institutions– as AAA/SWITCH in the project competition

– Stakeholders: Scientists, IT services

Page 29: Infrastructure and Basic Grid Services

2008 January 15 29Infrastructure and Basic Grid Services WG

Subgroups

Data Management:Can Tuerker, Wibke Sudholt, Peter Kunszt, Chad La Joie, Christoph Witzig, Alessandro Usai, Roberto Fabbrettihttp://www.doodle.ch/vgivau2akspintqx – Assemble requirements and initial design for DM services

Grid Operations Team GOT: Micheala Thiemard, Pascal Jermini, Wibke Sudholt, Christian Bolliger, Sigve Haug, Sergio Maffioletti, Placi Flury Alessandro Usai http://doodle.ch/ytpnscqvgyhrah47 – Define the initial operating infrastructure pragmatically today– Define operational procedures and processes as possible

Grid Architecture Team GAT: Chad LaJoie, Peter Kunszt, Wibke Sudholt, Can Tuerker, Derek Feichtinger, Christoph Witzig http://doodle.ch/pyp4s64vbdsqaisc – Define the future Grid architecture for SwiNG