67,432 research outputs found
The Orbital Space Environment and Space Situational Awareness Domain Ontology â Towards an International Information System for Space Data
The orbital space environment is home to natural and artificial satellites, debris, and space weather phenomena. As the population of orbital objects grows so do the potential hazards to astronauts, space infrastructure and spaceflight capability. Orbital debris, in particular, is a universal concern. This and other hazards can be minimized by improving global space situational awareness (SSA). By sharing more data and increasing observational coverage of the space environment we stand to achieve that goal, thereby making spaceflight safer and expanding our knowledge of near-Earth space. To facilitate data-sharing interoperability among distinct orbital debris and space object catalogs, and SSA information systems, I proposed ontology in (Rovetto, 2015) and (Rovetto and Kelso, 2016). I continue this effort toward formal representations and models of the overall domain that may serve to improve peaceful SSA and increase our scientific knowledge.
This paper explains the project concept introduced in those publications, summarizing efforts to date as well as the research field of ontology development and engineering. I describe concepts for an ontological framework for the orbital space environment, near-Earth space environment and SSA domain. An ontological framework is conceived as a part of a potential international information system. The purpose of such a system is to consolidate, analyze and reason over various sources and types of orbital and SSA data toward the mutually beneficial goals of safer space navigation and scientific research. Recent internationals findings on the limitations of orbital data, in addition to existing publications on collaborative SSA, demonstrate both the overlap with this project and the need for data-sharing and integration
Optical Synoptic Telescopes: New Science Frontiers
Over the past decade, sky surveys such as the Sloan Digital Sky Survey have
proven the power of large data sets for answering fundamental astrophysical
questions. This observational progress, based on a synergy of advances in
telescope construction, detectors, and information technology, has had a
dramatic impact on nearly all fields of astronomy, and areas of fundamental
physics. The next-generation instruments, and the surveys that will be made
with them, will maintain this revolutionary progress. The hardware and
computational technical challenges and the exciting science opportunities are
attracting scientists and engineers from astronomy, optics, low-light-level
detectors, high-energy physics, statistics, and computer science. The history
of astronomy has taught us repeatedly that there are surprises whenever we view
the sky in a new way. This will be particularly true of discoveries emerging
from a new generation of sky surveys. Imaging data from large ground-based
active optics telescopes with sufficient etendue can address many scientific
missions simultaneously. These new investigations will rely on the statistical
precision obtainable with billions of objects. For the first time, the full sky
will be surveyed deep and fast, opening a new window on a universe of faint
moving and distant exploding objects as well as unraveling the mystery of dark
energy.Comment: 12 pages, 7 figure
A grid-based infrastructure for distributed retrieval
In large-scale distributed retrieval, challenges of latency, heterogeneity, and dynamicity emphasise the importance of infrastructural support in reducing the development costs of state-of-the-art solutions. We present a service-based infrastructure for distributed retrieval which blends middleware facilities and a design framework to âliftâ the resource sharing approach and the computational services of a European Grid platform into the domain of e-Science applications. In this paper, we give an overview of the DILIGENT Search Framework and illustrate its exploitation in the ïŹeld of Earth Science
Building a Disciplinary, World-Wide Data Infrastructure
Sharing scientific data, with the objective of making it fully discoverable,
accessible, assessable, intelligible, usable, and interoperable, requires work
at the disciplinary level to define in particular how the data should be
formatted and described. Each discipline has its own organization and history
as a starting point, and this paper explores the way a range of disciplines,
namely materials science, crystallography, astronomy, earth sciences,
humanities and linguistics get organized at the international level to tackle
this question. In each case, the disciplinary culture with respect to data
sharing, science drivers, organization and lessons learnt are briefly
described, as well as the elements of the specific data infrastructure which
are or could be shared with others. Commonalities and differences are assessed.
Common key elements for success are identified: data sharing should be science
driven; defining the disciplinary part of the interdisciplinary standards is
mandatory but challenging; sharing of applications should accompany data
sharing. Incentives such as journal and funding agency requirements are also
similar. For all, it also appears that social aspects are more challenging than
technological ones. Governance is more diverse, and linked to the discipline
organization. CODATA, the RDA and the WDS can facilitate the establishment of
disciplinary interoperability frameworks. Being problem-driven is also a key
factor of success for building bridges to enable interdisciplinary research.Comment: Proceedings of the session "Building a disciplinary, world-wide data
infrastructure" of SciDataCon 2016, held in Denver, CO, USA, 12-14 September
2016, to be published in ICSU CODATA Data Science Journal in 201
Invest to Save: Report and Recommendations of the NSF-DELOS Working Group on Digital Archiving and Preservation
Digital archiving and preservation are important areas for research and development, but there is no agreed upon set of priorities or coherent plan for research in this area. Research projects in this area tend to be small and driven by particular institutional problems or concerns. As a consequence, proposed solutions from experimental projects and prototypes tend not to scale to millions of digital objects, nor do the results from disparate projects readily build on each other. It is also unclear whether it is worthwhile to seek general solutions or whether different strategies are needed for different types of digital objects and collections. The lack of coordination in both research and development means that there are some areas where researchers are reinventing the wheel while other areas are neglected.
Digital archiving and preservation is an area that will benefit from an exercise in analysis, priority setting, and planning for future research. The WG aims to survey current research activities, identify gaps, and develop a white paper proposing future research directions in the area of digital preservation. Some of the potential areas for research include repository architectures and inter-operability among digital archives; automated tools for capture, ingest, and normalization of digital objects; and harmonization of preservation formats and metadata. There can also be opportunities for development of commercial products in the areas of mass storage systems, repositories and repository management systems, and data management software and tools.
SpaceNet MVOI: a Multi-View Overhead Imagery Dataset
Detection and segmentation of objects in overheard imagery is a challenging
task. The variable density, random orientation, small size, and
instance-to-instance heterogeneity of objects in overhead imagery calls for
approaches distinct from existing models designed for natural scene datasets.
Though new overhead imagery datasets are being developed, they almost
universally comprise a single view taken from directly overhead ("at nadir"),
failing to address a critical variable: look angle. By contrast, views vary in
real-world overhead imagery, particularly in dynamic scenarios such as natural
disasters where first looks are often over 40 degrees off-nadir. This
represents an important challenge to computer vision methods, as changing view
angle adds distortions, alters resolution, and changes lighting. At present,
the impact of these perturbations for algorithmic detection and segmentation of
objects is untested. To address this problem, we present an open source
Multi-View Overhead Imagery dataset, termed SpaceNet MVOI, with 27 unique looks
from a broad range of viewing angles (-32.5 degrees to 54.0 degrees). Each of
these images cover the same 665 square km geographic extent and are annotated
with 126,747 building footprint labels, enabling direct assessment of the
impact of viewpoint perturbation on model performance. We benchmark multiple
leading segmentation and object detection models on: (1) building detection,
(2) generalization to unseen viewing angles and resolutions, and (3)
sensitivity of building footprint extraction to changes in resolution. We find
that state of the art segmentation and object detection models struggle to
identify buildings in off-nadir imagery and generalize poorly to unseen views,
presenting an important benchmark to explore the broadly relevant challenge of
detecting small, heterogeneous target objects in visually dynamic contexts.Comment: Accepted into IEEE International Conference on Computer Vision (ICCV)
201
CERN openlab Whitepaper on Future IT Challenges in Scientific Research
This whitepaper describes the major IT challenges in scientific research at CERN and several other European and international research laboratories and projects. Each challenge is exemplified through a set of concrete use cases drawn from the requirements of large-scale scientific programs. The paper is based on contributions from many researchers and IT experts of the participating laboratories and also input from the existing CERN openlab industrial sponsors. The views expressed in this document are those of the individual contributors and do not necessarily reflect the view of their organisations and/or affiliates
Theory and Practice of Data Citation
Citations are the cornerstone of knowledge propagation and the primary means
of assessing the quality of research, as well as directing investments in
science. Science is increasingly becoming "data-intensive", where large volumes
of data are collected and analyzed to discover complex patterns through
simulations and experiments, and most scientific reference works have been
replaced by online curated datasets. Yet, given a dataset, there is no
quantitative, consistent and established way of knowing how it has been used
over time, who contributed to its curation, what results have been yielded or
what value it has.
The development of a theory and practice of data citation is fundamental for
considering data as first-class research objects with the same relevance and
centrality of traditional scientific products. Many works in recent years have
discussed data citation from different viewpoints: illustrating why data
citation is needed, defining the principles and outlining recommendations for
data citation systems, and providing computational methods for addressing
specific issues of data citation.
The current panorama is many-faceted and an overall view that brings together
diverse aspects of this topic is still missing. Therefore, this paper aims to
describe the lay of the land for data citation, both from the theoretical (the
why and what) and the practical (the how) angle.Comment: 24 pages, 2 tables, pre-print accepted in Journal of the Association
for Information Science and Technology (JASIST), 201
Data fluidity in DARIAH -- pushing the agenda forward
This paper provides both an update concerning the setting up of the European
DARIAH infrastructure and a series of strong action lines related to the
development of a data centred strategy for the humanities in the coming years.
In particular we tackle various aspect of data management: data hosting, the
setting up of a DARIAH seal of approval, the establishment of a charter between
cultural heritage institutions and scholars and finally a specific view on
certification mechanisms for data
Citation and peer review of data: moving towards formal data publication
This paper discusses many of the issues associated with formally publishing data in academia, focusing primarily on the structures that need to be put in place for peer review and formal citation of datasets. Data publication is becoming increasingly important to the scientific community, as it will provide a mechanism for those who create data to receive academic credit for their work and will allow the conclusions arising from an analysis to be more readily verifiable, thus promoting transparency in the scientific process. Peer review of data will also provide a mechanism for ensuring the quality of datasets, and we provide suggestions on the types of activities one expects to see in the peer review of data. A simple taxonomy of data publication methodologies is presented and evaluated, and the paper concludes with a discussion of dataset granularity, transience and semantics, along with a recommended human-readable citation syntax
- âŠ