11 research outputs found

    Regularization and semi-supervised learning on large graphs

    No full text
    Abstract. We consider the problem of labeling a partially labeled graph. This setting may arise in a number of situations from survey sampling to information retrieval to pattern recognition in manifold settings. It is also of potential practical importance, when the data is abundant, but labeling is expensive or requires human assistance. Our approach develops a framework for regularization on such graphs. The algorithms are very simple and involve solving a single, usually sparse, system of linear equations. Using the notion of algorithmic stability, we derive bounds on the generalization error and relate it to structural invariants of the graph. Some experimental results testing the performance of the regularization algorithm and the usefulness of the generalization bound are presented.

    Characterization of the probabilistic traveling salesman problem

    Get PDF
    We show that stochastic annealing can be successfully applied to gain new results on the probabilistic traveling salesman problem. The probabilistic "traveling salesman" must decide on an a priori order in which to visit n cities (randomly distributed over a unit square) before learning that some cities can be omitted. We find the optimized average length of the pruned tour follows E((L) over bar (pruned))=rootnp(0.872-0.105p)f(np), where p is the probability of a city needing to be visited, and f(np)-->1 as np-->infinity. The average length of the a priori tour (before omitting any cities) is found to follow E(L-a priori)=rootn/pbeta(p), where beta(p)=1/[1.25-0.82 ln(p)] is measured for 0.05less than or equal topless than or equal to0.6. Scaling arguments and indirect measurements suggest that beta(p) tends towards a constant for p<0.03. Our stochastic annealing algorithm is based on limited sampling of the pruned tour lengths, exploiting the sampling error to provide the analog of thermal fluctuations in simulated (thermal) annealing. The method has general application to the optimization of functions whose cost to evaluate rises with the precision required
    corecore