39,373 research outputs found
Metric Learning for Individual Fairness
There has been much discussion concerning how "fairness" should be measured or enforced in classification. Individual Fairness [Dwork et al., 2012], which requires that similar individuals be treated similarly, is a highly appealing definition as it gives strong treatment guarantees for individuals. Unfortunately, the need for a task-specific similarity metric has prevented its use in practice. In this work, we propose a solution to the problem of approximating a metric for Individual Fairness based on human judgments. Our model assumes access to a human fairness arbiter who is free of explicit biases and possesses sufficient domain knowledge to evaluate similarity. Our contributions include definitions for metric approximation relevant for Individual Fairness, constructions for approximations from a limited number of realistic queries to the arbiter on a sample of individuals, and learning procedures to construct hypotheses for metric approximations which generalize to unseen samples under certain assumptions of learnability of distance threshold functions
An efficient time optimized scheme for progressive analytics in big data
Big data analytics is the key research subject for future data driven decision making applications. Due to the large amount of data, progressive analytics could provide an efficient way for querying big data clusters. Each cluster contains only a piece of the examined data. Continuous queries over these data sources require intelligent mechanisms to result the final outcome (query response) in the minimum time with the maximum performance. A Query Controller (QC) is responsible to manage continuous/sequential queries and return the final outcome to users or applications. In this paper, we propose a mechanism that can be adopted by the QC. The proposed mechanism is capable of managing partial results retrieved by a number of processors each one responsible for each cluster. Each processor executes a query over a specific cluster of data. Our mechanism adopts two sequential decision making models for handling the incoming partial results. The first model is based on a finite horizon time-optimized model and the second one is based on an infinite horizon optimally scheduled model. We provide mathematical formulations for solving the discussed problem and present simulation results. Through a large number of experiments, we reveal the advantages of the proposed models and give numerical results comparing them with a deterministic model. These results indicate that the proposed models can efficiently reduce the required time for returning the final outcome to the user/application while keeping the quality of the aggregated result at high levels
Progressive Simplification of Polygonal Curves
Simplifying polygonal curves at different levels of detail is an important
problem with many applications. Existing geometric optimization algorithms are
only capable of minimizing the complexity of a simplified curve for a single
level of detail. We present an -time algorithm that takes a polygonal
curve of n vertices and produces a set of consistent simplifications for m
scales while minimizing the cumulative simplification complexity. This
algorithm is compatible with distance measures such as the Hausdorff, the
Fr\'echet and area-based distances, and enables simplification for continuous
scaling in time. To speed up this algorithm in practice, we present
new techniques for constructing and representing so-called shortcut graphs.
Experimental evaluation of these techniques on trajectory data reveals a
significant improvement of using shortcut graphs for progressive and
non-progressive curve simplification, both in terms of running time and memory
usage.Comment: 20 pages, 20 figure
- …