418 research outputs found
A perspective on the Healthgrid initiative
This paper presents a perspective on the Healthgrid initiative which involves
European projects deploying pioneering applications of grid technology in the
health sector. In the last couple of years, several grid projects have been
funded on health related issues at national and European levels. A crucial
issue is to maximize their cross fertilization in the context of an environment
where data of medical interest can be stored and made easily available to the
different actors in healthcare, physicians, healthcare centres and
administrations, and of course the citizens. The Healthgrid initiative,
represented by the Healthgrid association (http://www.healthgrid.org), was
initiated to bring the necessary long term continuity, to reinforce and promote
awareness of the possibilities and advantages linked to the deployment of GRID
technologies in health. Technologies to address the specific requirements for
medical applications are under development. Results from the DataGrid and other
projects are given as examples of early applications.Comment: 6 pages, 1 figure. Accepted by the Second International Workshop on
Biomedical Computations on the Grid, at the 4th IEEE/ACM International
Symposium on Cluster Computing and the Grid (CCGrid 2004). Chicago USA, April
200
Next-Generation EU DataGrid Data Management Services
We describe the architecture and initial implementation of the
next-generation of Grid Data Management Middleware in the EU DataGrid (EDG)
project.
The new architecture stems out of our experience and the users requirements
gathered during the two years of running our initial set of Grid Data
Management Services. All of our new services are based on the Web Service
technology paradigm, very much in line with the emerging Open Grid Services
Architecture (OGSA). We have modularized our components and invested a great
amount of effort towards a secure, extensible and robust service, starting from
the design but also using a streamlined build and testing framework.
Our service components are: Replica Location Service, Replica Metadata
Service, Replica Optimization Service, Replica Subscription and high-level
replica management. The service security infrastructure is fully GSI-enabled,
hence compatible with the existing Globus Toolkit 2-based services; moreover,
it allows for fine-grained authorization mechanisms that can be adjusted
depending on the service semantics.Comment: Talk from the 2003 Computing in High Energy and Nuclear Physics
(CHEP03), La Jolla,Ca, USA, March 2003 8 pages, LaTeX, the file contains all
LaTeX sources - figures are in the directory "figures
HEP Applications Evaluation of the EDG Testbed and Middleware
Workpackage 8 of the European Datagrid project was formed in January 2001
with representatives from the four LHC experiments, and with experiment
independent people from five of the six main EDG partners. In September 2002
WP8 was strengthened by the addition of effort from BaBar and D0. The original
mandate of WP8 was, following the definition of short- and long-term
requirements, to port experiment software to the EDG middleware and testbed
environment. A major additional activity has been testing the basic
functionality and performance of this environment. This paper reviews
experiences and evaluations in the areas of job submission, data management,
mass storage handling, information systems and monitoring. It also comments on
the problems of remote debugging, the portability of code, and scaling problems
with increasing numbers of jobs, sites and nodes. Reference is made to the
pioneeering work of Atlas and CMS in integrating the use of the EDG Testbed
into their data challenges. A forward look is made to essential software
developments within EDG and to the necessary cooperation between EDG and LCG
for the LCG prototype due in mid 2003.Comment: Talk from the 2003 Computing in High Energy and Nuclear Physics
Conference (CHEP03), La Jolla, CA, USA, March 2003, 7 pages. PSN THCT00
An Overview of a Grid Architecture for Scientific Computing
This document gives an overview of a Grid testbed architecture proposal for
the NorduGrid project. The aim of the project is to establish an inter-Nordic
testbed facility for implementation of wide area computing and data handling.
The architecture is supposed to define a Grid system suitable for solving data
intensive problems at the Large Hadron Collider at CERN. We present the various
architecture components needed for such a system. After that we go on to give a
description of the dynamics by showing the task flow
The NorduGrid architecture and tools
The NorduGrid project designed a Grid architecture with the primary goal to
meet the requirements of production tasks of the LHC experiments. While it is
meant to be a rather generic Grid system, it puts emphasis on batch processing
suitable for problems encountered in High Energy Physics. The NorduGrid
architecture implementation uses the \globus{} as the foundation for various
components, developed by the project. While introducing new services, the
NorduGrid does not modify the Globus tools, such that the two can eventually
co-exist. The NorduGrid topology is decentralized, avoiding a single point of
failure. The NorduGrid architecture is thus a light-weight, non-invasive and
dynamic one, while robust and scalable, capable of meeting most challenging
tasks of High Energy Physics.Comment: Talk from the 2003 Computing in High Energy Physics and Nuclear
Physics (CHEP03), La Jolla, Ca, USA, March 2003, 9 pages,LaTeX, 4 figures.
PSN MOAT00
The INFN-grid testbed
The Italian INFN-Grid Project is committed to set-up, run and manage an unprecedented nation-wide Grid infrastructure. The implementation and use of this INFN-Grid Testbed is presented and discussed. Particular care and attention are devoted to those activities, relevant for the management of the Testbed, carried out by the INFN within international Grid Projects
Grid Computing: Concepts and Applications
The challenge of CERN experiments at the Large Hadron Collider (LHC), which will collect data at rates in the range of PBs/year, requires the development of GRID technologies to optimize the exploitation of distributed computing power and the automatic access to distributed data storage. Several projects are addressing the problem of setting up the hardware infrastructure of a GRID, as well as the development of the middleware required to manage it: a working GRID should look like a set of services, accessible to registered applications, which will help cooperate the different computing and storage resources. As it happened for the World Wide Web, GRID concepts are in principle important not only for High Energy Physics (HEP): for this reason, GRID developers, while keeping in mind the needs of HEP experiments, are trying to design GRID services in the most general way. As examples, two applications are described: the CERN/ALICE experiment at the LHC and a recently approved INFN project (GPCALMA) which will set up a GRID prototype between several mammographic centres in Italy
Installing, Running and Maintaining Large Linux Clusters at CERN
Having built up Linux clusters to more than 1000 nodes over the past five
years, we already have practical experience confronting some of the LHC scale
computing challenges: scalability, automation, hardware diversity, security,
and rolling OS upgrades. This paper describes the tools and processes we have
implemented, working in close collaboration with the EDG project [1],
especially with the WP4 subtask, to improve the manageability of our clusters,
in particular in the areas of system installation, configuration, and
monitoring. In addition to the purely technical issues, providing shared
interactive and batch services which can adapt to meet the diverse and changing
requirements of our users is a significant challenge. We describe the
developments and tuning that we have introduced on our LSF based systems to
maximise both responsiveness to users and overall system utilisation. Finally,
this paper will describe the problems we are facing in enlarging our
heterogeneous Linux clusters, the progress we have made in dealing with the
current issues and the steps we are taking to gridify the clustersComment: 5 pages, Proceedings for the CHEP 2003 conference, La Jolla,
California, March 24 - 28, 200
Grid technology for biomedical applications
International audienceThe deployment of biomedical applications in a grid environment has started about three years ago in several European projects and national ini-tiatives. These applications have demonstrated that the grid paradigm was rele-vant to the needs of the biomedical community. They have also highlighted that this community had very specific requirements on middleware and needed fur-ther structuring in large collaborations in order to participate to the deployment of grid infrastructures in the coming years. In this paper, we propose several ar-eas where grid technology can today improve research and healthcare. A cru-cial issue is to maximize the cross fertilization among projects in the perspec-tive of an environment where data of medical interest can be stored and made easily available to the different actors of healthcare, the physicians, the health-care centres and administrations, and of course the citizens
- …