27,949 research outputs found
DeepBrain: Functional Representation of Neural In-Situ Hybridization Images for Gene Ontology Classification Using Deep Convolutional Autoencoders
This paper presents a novel deep learning-based method for learning a
functional representation of mammalian neural images. The method uses a deep
convolutional denoising autoencoder (CDAE) for generating an invariant, compact
representation of in situ hybridization (ISH) images. While most existing
methods for bio-imaging analysis were not developed to handle images with
highly complex anatomical structures, the results presented in this paper show
that functional representation extracted by CDAE can help learn features of
functional gene ontology categories for their classification in a highly
accurate manner. Using this CDAE representation, our method outperforms the
previous state-of-the-art classification rate, by improving the average AUC
from 0.92 to 0.98, i.e., achieving 75% reduction in error. The method operates
on input images that were downsampled significantly with respect to the
original ones to make it computationally feasible
Knowledge will Propel Machine Understanding of Content: Extrapolating from Current Examples
Machine Learning has been a big success story during the AI resurgence. One
particular stand out success relates to learning from a massive amount of data.
In spite of early assertions of the unreasonable effectiveness of data, there
is increasing recognition for utilizing knowledge whenever it is available or
can be created purposefully. In this paper, we discuss the indispensable role
of knowledge for deeper understanding of content where (i) large amounts of
training data are unavailable, (ii) the objects to be recognized are complex,
(e.g., implicit entities and highly subjective content), and (iii) applications
need to use complementary or related data in multiple modalities/media. What
brings us to the cusp of rapid progress is our ability to (a) create relevant
and reliable knowledge and (b) carefully exploit knowledge to enhance ML/NLP
techniques. Using diverse examples, we seek to foretell unprecedented progress
in our ability for deeper understanding and exploitation of multimodal data and
continued incorporation of knowledge in learning techniques.Comment: Pre-print of the paper accepted at 2017 IEEE/WIC/ACM International
Conference on Web Intelligence (WI). arXiv admin note: substantial text
overlap with arXiv:1610.0770
TAIP: an anytime algorithm for allocating student teams to internship programs
In scenarios that require teamwork, we usually have at hand a variety of
specific tasks, for which we need to form a team in order to carry out each
one. Here we target the problem of matching teams with tasks within the context
of education, and specifically in the context of forming teams of students and
allocating them to internship programs. First we provide a formalization of the
Team Allocation for Internship Programs Problem, and show the computational
hardness of solving it optimally. Thereafter, we propose TAIP, a heuristic
algorithm that generates an initial team allocation which later on attempts to
improve in an iterative process. Moreover, we conduct a systematic evaluation
to show that TAIP reaches optimality, and outperforms CPLEX in terms of time.Comment: 10 pages, 7 figure
Identification of Design Principles
This report identifies those design principles for a (possibly new) query and transformation
language for the Web supporting inference that are considered essential. Based upon these
design principles an initial strawman is selected. Scenarios for querying the Semantic Web
illustrate the design principles and their reflection in the initial strawman, i.e., a first draft of
the query language to be designed and implemented by the REWERSE working group I4
Grand Challenges of Traceability: The Next Ten Years
In 2007, the software and systems traceability community met at the first
Natural Bridge symposium on the Grand Challenges of Traceability to establish
and address research goals for achieving effective, trustworthy, and ubiquitous
traceability. Ten years later, in 2017, the community came together to evaluate
a decade of progress towards achieving these goals. These proceedings document
some of that progress. They include a series of short position papers,
representing current work in the community organized across four process axes
of traceability practice. The sessions covered topics from Trace Strategizing,
Trace Link Creation and Evolution, Trace Link Usage, real-world applications of
Traceability, and Traceability Datasets and benchmarks. Two breakout groups
focused on the importance of creating and sharing traceability datasets within
the research community, and discussed challenges related to the adoption of
tracing techniques in industrial practice. Members of the research community
are engaged in many active, ongoing, and impactful research projects. Our hope
is that ten years from now we will be able to look back at a productive decade
of research and claim that we have achieved the overarching Grand Challenge of
Traceability, which seeks for traceability to be always present, built into the
engineering process, and for it to have "effectively disappeared without a
trace". We hope that others will see the potential that traceability has for
empowering software and systems engineers to develop higher-quality products at
increasing levels of complexity and scale, and that they will join the active
community of Software and Systems traceability researchers as we move forward
into the next decade of research
- …