8,795 research outputs found
Real-time Planning as Decision-making Under Uncertainty
In real-time planning, an agent must select the next action to take within a fixed time bound.
Many popular real-time heuristic search methods approach this by expanding nodes using time-limited A* and selecting the action leading toward the frontier node with the lowest f value. In this thesis, we reconsider real-time planning as a problem of decision-making under uncertainty. We treat heuristic values as uncertain evidence and we explore several backup methods for aggregating this evidence. We then propose a novel lookahead strategy that expands nodes to minimize risk, the expected regret in case a non-optimal action is chosen. We evaluate these methods in a simple synthetic benchmark and the sliding tile puzzle and find that they outperform previous methods. This work illustrates how uncertainty can arise even when solving deterministic planning problems, due to the inherent ignorance of time-limited search algorithms about those portions of the state space that they have not computed, and how an agent can benefit from explicitly meta-reasoning about this uncertainty
FindFoci: a focus detection algorithm with automated parameter training that closely matches human assignments, reduces human inconsistencies and increases speed of analysis
Accurate and reproducible quantification of the accumulation of proteins into foci in cells is essential for data interpretation and for biological inferences. To improve reproducibility, much emphasis has been placed on the preparation of samples, but less attention has been given to reporting and standardizing the quantification of foci. The current standard to quantitate foci in open-source software is to manually determine a range of parameters based on the outcome of one or a few representative images and then apply the parameter combination to the analysis of a larger dataset. Here, we demonstrate the power and utility of using machine learning to train a new algorithm (FindFoci) to determine optimal parameters. FindFoci closely matches human assignments and allows rapid automated exploration of parameter space. Thus, individuals can train the algorithm to mirror their own assignments and then automate focus counting using the same parameters across a large number of images. Using the training algorithm to match human assignments of foci, we demonstrate that applying an optimal parameter combination from a single image is not broadly applicable to analysis of other images scored by the same experimenter or by other experimenters. Our analysis thus reveals wide variation in human assignment of foci and their quantification. To overcome this, we developed training on multiple images, which reduces the inconsistency of using a single or a few images to set parameters for focus detection. FindFoci is provided as an open-source plugin for ImageJ
Submodular relaxation for inference in Markov random fields
In this paper we address the problem of finding the most probable state of a
discrete Markov random field (MRF), also known as the MRF energy minimization
problem. The task is known to be NP-hard in general and its practical
importance motivates numerous approximate algorithms. We propose a submodular
relaxation approach (SMR) based on a Lagrangian relaxation of the initial
problem. Unlike the dual decomposition approach of Komodakis et al., 2011 SMR
does not decompose the graph structure of the initial problem but constructs a
submodular energy that is minimized within the Lagrangian relaxation. Our
approach is applicable to both pairwise and high-order MRFs and allows to take
into account global potentials of certain types. We study theoretical
properties of the proposed approach and evaluate it experimentally.Comment: This paper is accepted for publication in IEEE Transactions on
Pattern Analysis and Machine Intelligenc
Using Machine Learning to Predict the Evolution of Physics Research
The advancement of science as outlined by Popper and Kuhn is largely
qualitative, but with bibliometric data it is possible and desirable to develop
a quantitative picture of scientific progress. Furthermore it is also important
to allocate finite resources to research topics that have growth potential, to
accelerate the process from scientific breakthroughs to technological
innovations. In this paper, we address this problem of quantitative knowledge
evolution by analysing the APS publication data set from 1981 to 2010. We build
the bibliographic coupling and co-citation networks, use the Louvain method to
detect topical clusters (TCs) in each year, measure the similarity of TCs in
consecutive years, and visualize the results as alluvial diagrams. Having the
predictive features describing a given TC and its known evolution in the next
year, we can train a machine learning model to predict future changes of TCs,
i.e., their continuing, dissolving, merging and splitting. We found the number
of papers from certain journals, the degree, closeness, and betweenness to be
the most predictive features. Additionally, betweenness increases significantly
for merging events, and decreases significantly for splitting events. Our
results represent a first step from a descriptive understanding of the Science
of Science (SciSci), towards one that is ultimately prescriptive.Comment: 24 pages, 10 figures, 4 tables, supplementary information is include
Normalized Affymetrix expression data are biased by G-quadruplex formation
Probes with runs of four or more guanines (G-stacks) in their sequences can exhibit a level of hybridization that is unrelated to the expression levels of the mRNA that they are intended to measure. This is most likely caused by the formation of G-quadruplexes, where inter-probe guanines form Hoogsteen hydrogen bonds, which probes with G-stacks are capable of forming. We demonstrate that for a specific microarray data set using the Human HG-U133A Affymetrix GeneChip and RMA normalization there is significant bias in the expression levels, the fold change and the correlations between expression levels. These effects grow more pronounced as the number of G-stack probes in a probe set increases. Approximately 14 of the probe sets are directly affected. The analysis was repeated for a number of other normalization pipelines and two, FARMS and PLIER, minimized the bias to some extent. We estimate that ∼15 of the data sets deposited in the GEO database are susceptible to the effect. The inclusion of G-stack probes in the affected data sets can bias key parameters used in the selection and clustering of genes. The elimination of these probes from any analysis in such affected data sets outweighs the increase of noise in the signal. © 2011 The Author(s)
QoE-Based Low-Delay Live Streaming Using Throughput Predictions
Recently, HTTP-based adaptive streaming has become the de facto standard for
video streaming over the Internet. It allows clients to dynamically adapt media
characteristics to network conditions in order to ensure a high quality of
experience, that is, minimize playback interruptions, while maximizing video
quality at a reasonable level of quality changes. In the case of live
streaming, this task becomes particularly challenging due to the latency
constraints. The challenge further increases if a client uses a wireless
network, where the throughput is subject to considerable fluctuations.
Consequently, live streams often exhibit latencies of up to 30 seconds. In the
present work, we introduce an adaptation algorithm for HTTP-based live
streaming called LOLYPOP (Low-Latency Prediction-Based Adaptation) that is
designed to operate with a transport latency of few seconds. To reach this
goal, LOLYPOP leverages TCP throughput predictions on multiple time scales,
from 1 to 10 seconds, along with an estimate of the prediction error
distribution. In addition to satisfying the latency constraint, the algorithm
heuristically maximizes the quality of experience by maximizing the average
video quality as a function of the number of skipped segments and quality
transitions. In order to select an efficient prediction method, we studied the
performance of several time series prediction methods in IEEE 802.11 wireless
access networks. We evaluated LOLYPOP under a large set of experimental
conditions limiting the transport latency to 3 seconds, against a
state-of-the-art adaptation algorithm from the literature, called FESTIVE. We
observed that the average video quality is by up to a factor of 3 higher than
with FESTIVE. We also observed that LOLYPOP is able to reach a broader region
in the quality of experience space, and thus it is better adjustable to the
user profile or service provider requirements.Comment: Technical Report TKN-16-001, Telecommunication Networks Group,
Technische Universitaet Berlin. This TR updated TR TKN-15-00
- …