13,221 research outputs found
Interoperability in the OpenDreamKit Project: The Math-in-the-Middle Approach
OpenDreamKit --- "Open Digital Research Environment Toolkit for the
Advancement of Mathematics" --- is an H2020 EU Research Infrastructure project
that aims at supporting, over the period 2015--2019, the ecosystem of
open-source mathematical software systems. From that, OpenDreamKit will deliver
a flexible toolkit enabling research groups to set up Virtual Research
Environments, customised to meet the varied needs of research projects in pure
mathematics and applications.
An important step in the OpenDreamKit endeavor is to foster the
interoperability between a variety of systems, ranging from computer algebra
systems over mathematical databases to front-ends. This is the mission of the
integration work package (WP6). We report on experiments and future plans with
the \emph{Math-in-the-Middle} approach. This information architecture consists
in a central mathematical ontology that documents the domain and fixes a joint
vocabulary, combined with specifications of the functionalities of the various
systems. Interaction between systems can then be enriched by pivoting off this
information architecture.Comment: 15 pages, 7 figure
The VVV Templates Project. Towards an Automated Classification of VVV Light-Curves. I. Building a database of stellar variability in the near-infrared
Context. The Vista Variables in the V\'ia L\'actea (VVV) ESO Public Survey is
a variability survey of the Milky Way bulge and an adjacent section of the disk
carried out from 2010 on ESO Visible and Infrared Survey Telescope for
Astronomy (VISTA). VVV will eventually deliver a deep near-IR atlas with
photometry and positions in five passbands (ZYJHK_S) and a catalogue of 1-10
million variable point sources - mostly unknown - which require
classifications. Aims. The main goal of the VVV Templates Project, that we
introduce in this work, is to develop and test the machine-learning algorithms
for the automated classification of the VVV light-curves. As VVV is the first
massive, multi-epoch survey of stellar variability in the near-infrared, the
template light-curves that are required for training the classification
algorithms are not available. In the first paper of the series we describe the
construction of this comprehensive database of infrared stellar variability.
Methods. First we performed a systematic search in the literature and public
data archives, second, we coordinated a worldwide observational campaign, and
third we exploited the VVV variability database itself on (optically)
well-known stars to gather high-quality infrared light-curves of several
hundreds of variable stars. Results. We have now collected a significant (and
still increasing) number of infrared template light-curves. This database will
be used as a training-set for the machine-learning algorithms that will
automatically classify the light-curves produced by VVV. The results of such an
automated classification will be covered in forthcoming papers of the series.Comment: 12 pages, 16 figures, 3 tables, accepted for publication in A&A. Most
of the data are now accessible through http://www.vvvtemplates.org
Second Screen User Profiling and Multi-level Smart Recommendations in the context of Social TVs
In the context of Social TV, the increasing popularity of first and second
screen users, interacting and posting content online, illustrates new business
opportunities and related technical challenges, in order to enrich user
experience on such environments. SAM (Socializing Around Media) project uses
Social Media-connected infrastructure to deal with the aforementioned
challenges, providing intelligent user context management models and mechanisms
capturing social patterns, to apply collaborative filtering techniques and
personalized recommendations towards this direction. This paper presents the
Context Management mechanism of SAM, running in a Social TV environment to
provide smart recommendations for first and second screen content. Work
presented is evaluated using real movie rating dataset found online, to
validate the SAM's approach in terms of effectiveness as well as efficiency.Comment: In: Wu TT., Gennari R., Huang YM., Xie H., Cao Y. (eds) Emerging
Technologies for Education. SETE 201
MapReduce is Good Enough? If All You Have is a Hammer, Throw Away Everything That's Not a Nail!
Hadoop is currently the large-scale data analysis "hammer" of choice, but
there exist classes of algorithms that aren't "nails", in the sense that they
are not particularly amenable to the MapReduce programming model. To address
this, researchers have proposed MapReduce extensions or alternative programming
models in which these algorithms can be elegantly expressed. This essay
espouses a very different position: that MapReduce is "good enough", and that
instead of trying to invent screwdrivers, we should simply get rid of
everything that's not a nail. To be more specific, much discussion in the
literature surrounds the fact that iterative algorithms are a poor fit for
MapReduce: the simple solution is to find alternative non-iterative algorithms
that solve the same problem. This essay captures my personal experiences as an
academic researcher as well as a software engineer in a "real-world" production
analytics environment. From this combined perspective I reflect on the current
state and future of "big data" research
A conceptual approach to gene expression analysis enhanced by visual analytics
The analysis of gene expression data is a complex task for biologists wishing to understand the role of genes in the formation of diseases such as cancer. Biologists need greater support when trying to discover, and comprehend, new relationships within their data. In this paper, we describe an approach to the analysis of gene expression data where overlapping groupings are generated by Formal Concept Analysis and interactively analyzed in a tool called CUBIST. The CUBIST workflow involves querying a semantic database and converting the result into a formal context, which can be simplified to make it manageable, before it is visualized as a concept lattice and associated charts
An Introductory Guide to Aligning Networks Using SANA, the Simulated Annealing Network Aligner.
Sequence alignment has had an enormous impact on our understanding of biology, evolution, and disease. The alignment of biological networks holds similar promise. Biological networks generally model interactions between biomolecules such as proteins, genes, metabolites, or mRNAs. There is strong evidence that the network topology-the "structure" of the network-is correlated with the functions performed, so that network topology can be used to help predict or understand function. However, unlike sequence comparison and alignment-which is an essentially solved problem-network comparison and alignment is an NP-complete problem for which heuristic algorithms must be used.Here we introduce SANA, the Simulated Annealing Network Aligner. SANA is one of many algorithms proposed for the arena of biological network alignment. In the context of global network alignment, SANA stands out for its speed, memory efficiency, ease-of-use, and flexibility in the arena of producing alignments between two or more networks. SANA produces better alignments in minutes on a laptop than most other algorithms can produce in hours or days of CPU time on large server-class machines. We walk the user through how to use SANA for several types of biomolecular networks
- …