18,606 research outputs found
Deterministic Graph Exploration with Advice
We consider the task of graph exploration. An -node graph has unlabeled
nodes, and all ports at any node of degree are arbitrarily numbered
. A mobile agent has to visit all nodes and stop. The exploration
time is the number of edge traversals. We consider the problem of how much
knowledge the agent has to have a priori, in order to explore the graph in a
given time, using a deterministic algorithm. This a priori information (advice)
is provided to the agent by an oracle, in the form of a binary string, whose
length is called the size of advice. We consider two types of oracles. The
instance oracle knows the entire instance of the exploration problem, i.e., the
port-numbered map of the graph and the starting node of the agent in this map.
The map oracle knows the port-numbered map of the graph but does not know the
starting node of the agent.
We first consider exploration in polynomial time, and determine the exact
minimum size of advice to achieve it. This size is ,
for both types of oracles.
When advice is large, there are two natural time thresholds:
for a map oracle, and for an instance oracle, that can be achieved
with sufficiently large advice. We show that, with a map oracle, time
cannot be improved in general, regardless of the size of advice.
We also show that the smallest size of advice to achieve this time is larger
than , for any .
For an instance oracle, advice of size is enough to achieve time
. We show that, with any advice of size , the time of
exploration must be at least , for any , and with any
advice of size , the time must be .
We also investigate minimum advice sufficient for fast exploration of
hamiltonian graphs
Hierarchical Subquery Evaluation for Active Learning on a Graph
To train good supervised and semi-supervised object classifiers, it is
critical that we not waste the time of the human experts who are providing the
training labels. Existing active learning strategies can have uneven
performance, being efficient on some datasets but wasteful on others, or
inconsistent just between runs on the same dataset. We propose perplexity based
graph construction and a new hierarchical subquery evaluation algorithm to
combat this variability, and to release the potential of Expected Error
Reduction.
Under some specific circumstances, Expected Error Reduction has been one of
the strongest-performing informativeness criteria for active learning. Until
now, it has also been prohibitively costly to compute for sizeable datasets. We
demonstrate our highly practical algorithm, comparing it to other active
learning measures on classification datasets that vary in sparsity,
dimensionality, and size. Our algorithm is consistent over multiple runs and
achieves high accuracy, while querying the human expert for labels at a
frequency that matches their desired time budget.Comment: CVPR 201
- …