4,424 research outputs found
The EnMAP user interface and user request scenarios
EnMAP (Environmental Mapping and Analysis Program) is a German hyperspectral satellite mission providing high quality hyperspectral image data on a timely and frequent basis. Main objective is to investigate a wide range of ecosystem parameters encompassing agriculture, forestry, soil and geological environments, coastal zones and inland waters. The EnMAP Ground Segment will be designed, implemented and operated by the German Aerospace Center (DLR). The Applied Remote Sensing Cluster (DFD) at DLR is responsible for the establishment of a user interface. This paper provides details on the concept, design and functionality of the EnMAP user interface and a first analysis about potential user scenarios.
The user interface consists of two online portals. The EnMAP portal (www.enmap.org) provides general EnMAP mission information. It is the central entry point for all international users interested to learn about the EnMAP mission, its objectives, status, data products and processing chains. The EnMAP Data Access Portal (EDAP) is the entry point for any EnMAP data requests and comprises a set of service functions offered for every registered user. The scientific user is able to task the EnMAP HSI for Earth observations by providing tasking parameters, such as area, temporal aspects and allowed tilt angle.
In the second part of that paper different user scenarios according to the previously explained tasking parameters are presented and discussed in terms of their feasibility for scientific projects. For that purpose, a prototype of the observation planning tool enabling visualization of different user request scenarios was developed. It can be shown, that the number of data takes in a certain period of time increases with the latitude of the observation area. Further, the observation area can differ with the tilt angle of the satellite. Such findings can be crucial for the planning of remote sensing based projects, especially for those investigating ecosystem gradients in the time domain
European ALMA operations: the interaction with and support to the users
The Atacama Large Millimetre/submillimetre Array (ALMA) is one of the largest
and most complicated observatories ever built. Constructing and operating an
observatory at high altitude (5000m) in a cost effective and safe manner, with
minimal effect on the environment creates interesting challenges. Since the
array will have to adapt quickly to prevailing weather conditions, ALMA will be
operated exclusively in service mode. By the time of full science operations,
the fundamental ALMA data product shall be calibrated, deconvolved data cubes
and images, but raw data and data reduction software will be made available to
users as well. User support is provided by the ALMA Regional Centres (ARCs)
located in Europe, North America and Japan. These ARCs constitute the interface
between the user community and the ALMA observatory in Chile. For European
users the European ARC is being set up as a cluster of nodes located throughout
Europe, with the main centre at the ESO Headquarters in Garching. The main
centre serves as the access portal and in synergy with the distributed network
of ARC nodes, the main aim of the ARC is to optimize the ALMA science output
and to fully exploit this unique and powerful facility. The aim of this article
is to introduce the process of proposing for observing time, subsequent
execution of the observations, obtaining and processing of the data in the ALMA
epoch. The complete end-to-end process of the ALMA data flow from the proposal
submission to the data delivery is described.Comment: 7 pages, three figure
Report of the user requirements and web based access for eResearch workshops
The User Requirements and Web Based Access for eResearch Workshop, organized jointly by NeSC and NCeSS, was held on 19 May 2006. The aim was to identify lessons learned from e-Science projects that would contribute to our capacity to make Grid infrastructures and tools usable and accessible for diverse user communities. Its focus was on providing an opportunity for a pragmatic discussion between e-Science end users
and tool builders in order to understand usability challenges, technological options, community-specific content and needs, and methodologies for design and development. We invited members of six UK e-Science projects and one US project, trying as far as
possible to pair a user and developer from each project in order to discuss their contrasting perspectives and experiences. Three breakout group sessions covered the
topics of user-developer relations, commodification, and functionality. There was also extensive post-meeting discussion, summarized here.
Additional information on the workshop, including the agenda, participant list, and talk slides, can be found online at http://www.nesc.ac.uk/esi/events/685/
Reference: NeSC report UKeS-2006-07 available from http://www.nesc.ac.uk/technical_papers/UKeS-2006-07.pd
Resource provisioning in Science Clouds: Requirements and challenges
Cloud computing has permeated into the information technology industry in the
last few years, and it is emerging nowadays in scientific environments. Science
user communities are demanding a broad range of computing power to satisfy the
needs of high-performance applications, such as local clusters,
high-performance computing systems, and computing grids. Different workloads
are needed from different computational models, and the cloud is already
considered as a promising paradigm. The scheduling and allocation of resources
is always a challenging matter in any form of computation and clouds are not an
exception. Science applications have unique features that differentiate their
workloads, hence, their requirements have to be taken into consideration to be
fulfilled when building a Science Cloud. This paper will discuss what are the
main scheduling and resource allocation challenges for any Infrastructure as a
Service provider supporting scientific applications
INDIGO-Datacloud: foundations and architectural description of a Platform as a Service oriented to scientific computing
Software Engineering.-- et al.In this paper we describe the architecture of a Platform as a Service (PaaS) oriented to computing and data analysis. In order to clarify the choices we made, we explain the features using practical examples, applied to several known usage patterns in the area of HEP computing. The proposed architecture is devised to provide researchers with a unified view of distributed computing infrastructures, focusing in facilitating seamless access. In this respect the Platform is able to profit from the most recent developments for computing and processing large amounts of data, and to
exploit current storage and preservation technologies, with the appropriate mechanisms to ensure security and privacy.INDIGO-DataCloud is co-founded by the Horizon 2020Framework Programme.Peer reviewe
The Ultralight project: the network as an integrated and managed resource for data-intensive science
Looks at the UltraLight project which treats the network interconnecting globally distributed data sets as a dynamic, configurable, and closely monitored resource to construct a next-generation system that can meet the high-energy physics community's data-processing, distribution, access, and analysis needs
A Priority-Based Admission Control Scheme for Commercial Web Servers
This paper investigates into the performance and load management of web servers that are deployed in commercial websites. Such websites offer various services such as flight/hotel booking, online banking, stock trading, and product purchases among others. Customers are increasingly relying on these round-the-clock services which are easier and (generally) cheaper to order. However, such an increasing number of customersâ requests makes a greater demand on the web servers. This leads to web serversâ overload and the consequential provisioning of inadequate level of service. This paper addresses these issues and proposes an admission control scheme which is based on the class-based priority scheme that classifies customerâs requests into different classes. The proposed scheme is formally specified using Î Î -calculus and is implemented as a Java-based prototype system. The prototype system is used to simulate the behaviour of commercial website servers and to evaluate their performance in terms of response time, throughput, arrival rate, and the percentage of dropped requests. Experimental results demonstrate that the proposed scheme significantly improves the performance of high priority requests but without causing adverse effects on low priority requests
- âŠ