4,424 research outputs found

    The EnMAP user interface and user request scenarios

    Get PDF
    EnMAP (Environmental Mapping and Analysis Program) is a German hyperspectral satellite mission providing high quality hyperspectral image data on a timely and frequent basis. Main objective is to investigate a wide range of ecosystem parameters encompassing agriculture, forestry, soil and geological environments, coastal zones and inland waters. The EnMAP Ground Segment will be designed, implemented and operated by the German Aerospace Center (DLR). The Applied Remote Sensing Cluster (DFD) at DLR is responsible for the establishment of a user interface. This paper provides details on the concept, design and functionality of the EnMAP user interface and a first analysis about potential user scenarios. The user interface consists of two online portals. The EnMAP portal (www.enmap.org) provides general EnMAP mission information. It is the central entry point for all international users interested to learn about the EnMAP mission, its objectives, status, data products and processing chains. The EnMAP Data Access Portal (EDAP) is the entry point for any EnMAP data requests and comprises a set of service functions offered for every registered user. The scientific user is able to task the EnMAP HSI for Earth observations by providing tasking parameters, such as area, temporal aspects and allowed tilt angle. In the second part of that paper different user scenarios according to the previously explained tasking parameters are presented and discussed in terms of their feasibility for scientific projects. For that purpose, a prototype of the observation planning tool enabling visualization of different user request scenarios was developed. It can be shown, that the number of data takes in a certain period of time increases with the latitude of the observation area. Further, the observation area can differ with the tilt angle of the satellite. Such findings can be crucial for the planning of remote sensing based projects, especially for those investigating ecosystem gradients in the time domain

    European ALMA operations: the interaction with and support to the users

    Full text link
    The Atacama Large Millimetre/submillimetre Array (ALMA) is one of the largest and most complicated observatories ever built. Constructing and operating an observatory at high altitude (5000m) in a cost effective and safe manner, with minimal effect on the environment creates interesting challenges. Since the array will have to adapt quickly to prevailing weather conditions, ALMA will be operated exclusively in service mode. By the time of full science operations, the fundamental ALMA data product shall be calibrated, deconvolved data cubes and images, but raw data and data reduction software will be made available to users as well. User support is provided by the ALMA Regional Centres (ARCs) located in Europe, North America and Japan. These ARCs constitute the interface between the user community and the ALMA observatory in Chile. For European users the European ARC is being set up as a cluster of nodes located throughout Europe, with the main centre at the ESO Headquarters in Garching. The main centre serves as the access portal and in synergy with the distributed network of ARC nodes, the main aim of the ARC is to optimize the ALMA science output and to fully exploit this unique and powerful facility. The aim of this article is to introduce the process of proposing for observing time, subsequent execution of the observations, obtaining and processing of the data in the ALMA epoch. The complete end-to-end process of the ALMA data flow from the proposal submission to the data delivery is described.Comment: 7 pages, three figure

    Report of the user requirements and web based access for eResearch workshops

    Get PDF
    The User Requirements and Web Based Access for eResearch Workshop, organized jointly by NeSC and NCeSS, was held on 19 May 2006. The aim was to identify lessons learned from e-Science projects that would contribute to our capacity to make Grid infrastructures and tools usable and accessible for diverse user communities. Its focus was on providing an opportunity for a pragmatic discussion between e-Science end users and tool builders in order to understand usability challenges, technological options, community-specific content and needs, and methodologies for design and development. We invited members of six UK e-Science projects and one US project, trying as far as possible to pair a user and developer from each project in order to discuss their contrasting perspectives and experiences. Three breakout group sessions covered the topics of user-developer relations, commodification, and functionality. There was also extensive post-meeting discussion, summarized here. Additional information on the workshop, including the agenda, participant list, and talk slides, can be found online at http://www.nesc.ac.uk/esi/events/685/ Reference: NeSC report UKeS-2006-07 available from http://www.nesc.ac.uk/technical_papers/UKeS-2006-07.pd

    Resource provisioning in Science Clouds: Requirements and challenges

    Full text link
    Cloud computing has permeated into the information technology industry in the last few years, and it is emerging nowadays in scientific environments. Science user communities are demanding a broad range of computing power to satisfy the needs of high-performance applications, such as local clusters, high-performance computing systems, and computing grids. Different workloads are needed from different computational models, and the cloud is already considered as a promising paradigm. The scheduling and allocation of resources is always a challenging matter in any form of computation and clouds are not an exception. Science applications have unique features that differentiate their workloads, hence, their requirements have to be taken into consideration to be fulfilled when building a Science Cloud. This paper will discuss what are the main scheduling and resource allocation challenges for any Infrastructure as a Service provider supporting scientific applications

    INDIGO-Datacloud: foundations and architectural description of a Platform as a Service oriented to scientific computing

    Get PDF
    Software Engineering.-- et al.In this paper we describe the architecture of a Platform as a Service (PaaS) oriented to computing and data analysis. In order to clarify the choices we made, we explain the features using practical examples, applied to several known usage patterns in the area of HEP computing. The proposed architecture is devised to provide researchers with a unified view of distributed computing infrastructures, focusing in facilitating seamless access. In this respect the Platform is able to profit from the most recent developments for computing and processing large amounts of data, and to exploit current storage and preservation technologies, with the appropriate mechanisms to ensure security and privacy.INDIGO-DataCloud is co-founded by the Horizon 2020Framework Programme.Peer reviewe

    The Ultralight project: the network as an integrated and managed resource for data-intensive science

    Get PDF
    Looks at the UltraLight project which treats the network interconnecting globally distributed data sets as a dynamic, configurable, and closely monitored resource to construct a next-generation system that can meet the high-energy physics community's data-processing, distribution, access, and analysis needs

    A Priority-Based Admission Control Scheme for Commercial Web Servers

    Get PDF
    This paper investigates into the performance and load management of web servers that are deployed in commercial websites. Such websites offer various services such as flight/hotel booking, online banking, stock trading, and product purchases among others. Customers are increasingly relying on these round-the-clock services which are easier and (generally) cheaper to order. However, such an increasing number of customers’ requests makes a greater demand on the web servers. This leads to web servers’ overload and the consequential provisioning of inadequate level of service. This paper addresses these issues and proposes an admission control scheme which is based on the class-based priority scheme that classifies customer’s requests into different classes. The proposed scheme is formally specified using ΠΠ-calculus and is implemented as a Java-based prototype system. The prototype system is used to simulate the behaviour of commercial website servers and to evaluate their performance in terms of response time, throughput, arrival rate, and the percentage of dropped requests. Experimental results demonstrate that the proposed scheme significantly improves the performance of high priority requests but without causing adverse effects on low priority requests
    • 

    corecore