25 research outputs found

    Learning vector quantization for proximity data

    Get PDF
    Hofmann D. Learning vector quantization for proximity data. Bielefeld: Universität Bielefeld; 2016.Prototype-based classifiers such as learning vector quantization (LVQ) often display intuitive and flexible classification and learning rules. However, classical techniques are restricted to vectorial data only, and hence not suited for more complex data structures. Therefore, a few extensions of diverse LVQ variants to more general data which are characterized based on pairwise similarities or dissimilarities only have been proposed recently in the literature. In this contribution, we propose a novel extension of LVQ to similarity data which is based on the kernelization of an underlying probabilistic model: kernel robust soft LVQ (KRSLVQ). Relying on the notion of a pseudo-Euclidean embedding of proximity data, we put this specific approach as well as existing alternatives into a general framework which characterizes different fundamental possibilities how to extend LVQ towards proximity data: the main characteristics are given by the choice of the cost function, the interface to the data in terms of similarities or dissimilarities, and the way in which optimization takes place. In particular the latter strategy highlights the difference of popular kernel approaches versus so-called relational approaches. While KRSLVQ and alternatives lead to state of the art results, these extensions have two drawbacks as compared to their vectorial counterparts: (i) a quadratic training complexity is encountered due to the dependency of the methods on the full proximity matrix; (ii) prototypes are no longer given by vectors but they are represented in terms of an implicit linear combination of data, i.e. interpretability of the prototypes is lost. We investigate different techniques to deal with these challenges: We consider a speed-up of training by means of low rank approximations of the Gram matrix by its Nyström approximation. In benchmarks, this strategy is successful if the considered data are intrinsically low-dimensional. We propose a quick check to efficiently test this property prior to training. We extend KRSLVQ by sparse approximations of the prototypes: instead of the full coefficient vectors, few exemplars which represent the prototypes can be directly inspected by practitioners in the same way as data. We compare different paradigms based on which to infer a sparse approximation: sparsity priors while training, geometric approaches including orthogonal matching pursuit and core techniques, and heuristic approximations based on the coefficients or proximities. We demonstrate the performance of these LVQ techniques for benchmark data, reaching state of the art results. We discuss the behavior of the methods to enhance performance and interpretability as concerns quality, sparsity, and representativity, and we propose different measures how to quantitatively evaluate the performance of the approaches. We would like to point out that we had the possibility to present our findings in international publication organs including three journal articles [6, 9, 2], four conference papers [8, 5, 7, 1] and two workshop contributions [4, 3]. References [1] A. Gisbrecht, D. Hofmann, and B. Hammer. Discriminative dimensionality reduction mappings. Advances in Intelligent Data Analysis, 7619: 126–138, 2012. [2] B. Hammer, D. Hofmann, F.-M. Schleif, and X. Zhu. Learning vector quantization for (dis-)similarities. Neurocomputing, 131: 43–51, 2014. [3] D. Hofmann. Sparse approximations for kernel robust soft lvq. Mittweida Workshop on Computational Intelligence, 2013. [4] D. Hofmann, A. Gisbrecht, and B. Hammer. Discriminative probabilistic prototype based models in kernel space. New Challenges in Neural Computation, TR Machine Learning Reports, 2012. [5] D. Hofmann, A. Gisbrecht, and B. Hammer. Efficient approximations of kernel robust soft lvq. Workshop on Self-Organizing Maps, 198: 183–192, 2012. [6] D. Hofmann, A. Gisbrecht, and B. Hammer. Efficient approximations of robust soft learning vector quantization for non-vectorial data. Neurocomputing, 147: 96–106, 2015. [7] D. Hofmann and B. Hammer. Kernel robust soft learning vector quantization. Artificial Neural Networks in Pattern Recognition, 7477: 14–23, 2012. [8] D. Hofmann and B. Hammer. Sparse approximations for kernel learning vector quantization. European Symposium on Artificial Neural Networks, 549–554, 2013. [9] D. Hofmann, F.-M. Schleif, B. Paaßen, and B. Hammer. Learning interpretable kernelized prototype-based models. Neurocomputing, 141: 84–96, 2014

    Ahlfors circle maps and total reality: from Riemann to Rohlin

    Full text link
    This is a prejudiced survey on the Ahlfors (extremal) function and the weaker {\it circle maps} (Garabedian-Schiffer's translation of "Kreisabbildung"), i.e. those (branched) maps effecting the conformal representation upon the disc of a {\it compact bordered Riemann surface}. The theory in question has some well-known intersection with real algebraic geometry, especially Klein's ortho-symmetric curves via the paradigm of {\it total reality}. This leads to a gallery of pictures quite pleasant to visit of which we have attempted to trace the simplest representatives. This drifted us toward some electrodynamic motions along real circuits of dividing curves perhaps reminiscent of Kepler's planetary motions along ellipses. The ultimate origin of circle maps is of course to be traced back to Riemann's Thesis 1851 as well as his 1857 Nachlass. Apart from an abrupt claim by Teichm\"uller 1941 that everything is to be found in Klein (what we failed to assess on printed evidence), the pivotal contribution belongs to Ahlfors 1950 supplying an existence-proof of circle maps, as well as an analysis of an allied function-theoretic extremal problem. Works by Yamada 1978--2001, Gouma 1998 and Coppens 2011 suggest sharper degree controls than available in Ahlfors' era. Accordingly, our partisan belief is that much remains to be clarified regarding the foundation and optimal control of Ahlfors circle maps. The game of sharp estimation may look narrow-minded "Absch\"atzungsmathematik" alike, yet the philosophical outcome is as usual to contemplate how conformal and algebraic geometry are fighting together for the soul of Riemann surfaces. A second part explores the connection with Hilbert's 16th as envisioned by Rohlin 1978.Comment: 675 pages, 199 figures; extended version of the former text (v.1) by including now Rohlin's theory (v.2

    Crafting chaos: computational design of contraptions with complex behaviour

    Get PDF
    The 2010s saw the democratisation of digital fabrication technologies. Although this phenomenon made fabrication more accessible, physical assemblies displaying a complex behaviour are still difficult to design. While many methods support the creation of complex shapes and assemblies, managing a complex behaviour is often assumed to be a tedious aspect of the design process. As a result, the complex parts of the behaviour are either deemed negligible (when possible) or managed directly by the software, without offering much fine-grained user control. This thesis argues that efficient methods can support designers seeking complex behaviours by increasing their level of control over these behaviours. To demonstrate this, I study two types of artistic devices that are particularly challenging to design: drawing machines, and chain reaction contraptions. These artefacts’ complex behaviour can change dramatically even as their components are moved by a small amount. The first case study aims to facilitate the exploration and progressive refinement of complex patterns generated by drawing machines under drawing-level user-defined constraints. The approach was evaluated with a user study, and several machines drawing the expected pattern were fabricated. In the second case study, I propose an algorithm to optimise the layout of complex chain reaction contraptions described by a causal graph of events in order to make them robust to uncertainty. Several machines optimised with this method were successfully assembled and run. This thesis makes the following contributions: (1) support complex behaviour specifications; (2) enable users to easily explore design variations that respect these specifications; and (3) optimise the layout of a physical assembly to maximise the probability of real-life success

    Collection of abstracts of the 24th European Workshop on Computational Geometry

    Get PDF
    International audienceThe 24th European Workshop on Computational Geomety (EuroCG'08) was held at INRIA Nancy - Grand Est & LORIA on March 18-20, 2008. The present collection of abstracts contains the 63 scientific contributions as well as three invited talks presented at the workshop

    Fundamentals

    Get PDF
    Volume 1 establishes the foundations of this new field. It goes through all the steps from data collection, their summary and clustering, to different aspects of resource-aware learning, i.e., hardware, memory, energy, and communication awareness. Machine learning methods are inspected with respect to resource requirements and how to enhance scalability on diverse computing architectures ranging from embedded systems to large computing clusters

    Fundamentals

    Get PDF
    Volume 1 establishes the foundations of this new field. It goes through all the steps from data collection, their summary and clustering, to different aspects of resource-aware learning, i.e., hardware, memory, energy, and communication awareness. Machine learning methods are inspected with respect to resource requirements and how to enhance scalability on diverse computing architectures ranging from embedded systems to large computing clusters

    Large bichromatic point sets admit empty monochromatic 4-gons

    No full text
    We consider a variation of a problem stated by Erd˝os and Szekeres in 1935 about the existence of a number fES(k) such that any set S of at least fES(k) points in general position in the plane has a subset of k points that are the vertices of a convex k-gon. In our setting the points of S are colored, and we say that a (not necessarily convex) spanned polygon is monochromatic if all its vertices have the same color. Moreover, a polygon is called empty if it does not contain any points of S in its interior. We show that any bichromatic set of n ≥ 5044 points in R2 in general position determines at least one empty, monochromatic quadrilateral (and thus linearly many).Postprint (published version
    corecore