1,540 research outputs found
The Cost of Uncertainty in Curing Epidemics
Motivated by the study of controlling (curing) epidemics, we consider the
spread of an SI process on a known graph, where we have a limited budget to use
to transition infected nodes back to the susceptible state (i.e., to cure
nodes). Recent work has demonstrated that under perfect and instantaneous
information (which nodes are/are not infected), the budget required for curing
a graph precisely depends on a combinatorial property called the CutWidth. We
show that this assumption is in fact necessary: even a minor degradation of
perfect information, e.g., a diagnostic test that is 99% accurate, drastically
alters the landscape. Infections that could previously be cured in sublinear
time now may require exponential time, or orderwise larger budget to cure. The
crux of the issue comes down to a tension not present in the full information
case: if a node is suspected (but not certain) to be infected, do we risk
wasting our budget to try to cure an uninfected node, or increase our certainty
by longer observation, at the risk that the infection spreads further? Our
results present fundamental, algorithm-independent bounds that tradeoff budget
required vs. uncertainty.Comment: 35 pages, 3 figure
Link Prediction by De-anonymization: How We Won the Kaggle Social Network Challenge
This paper describes the winning entry to the IJCNN 2011 Social Network
Challenge run by Kaggle.com. The goal of the contest was to promote research on
real-world link prediction, and the dataset was a graph obtained by crawling
the popular Flickr social photo sharing website, with user identities scrubbed.
By de-anonymizing much of the competition test set using our own Flickr crawl,
we were able to effectively game the competition. Our attack represents a new
application of de-anonymization to gaming machine learning contests, suggesting
changes in how future competitions should be run.
We introduce a new simulated annealing-based weighted graph matching
algorithm for the seeding step of de-anonymization. We also show how to combine
de-anonymization with link prediction---the latter is required to achieve good
performance on the portion of the test set not de-anonymized---for example by
training the predictor on the de-anonymized portion of the test set, and
combining probabilistic predictions from de-anonymization and link prediction.Comment: 11 pages, 13 figures; submitted to IJCNN'201
Data based identification and prediction of nonlinear and complex dynamical systems
We thank Dr. R. Yang (formerly at ASU), Dr. R.-Q. Su (formerly at ASU), and Mr. Zhesi Shen for their contributions to a number of original papers on which this Review is partly based. This work was supported by ARO under Grant No. W911NF-14-1-0504. W.-X. Wang was also supported by NSFC under Grants No. 61573064 and No. 61074116, as well as by the Fundamental Research Funds for the Central Universities, Beijing Nova Programme.Peer reviewedPostprin
- …