301 research outputs found

    Geometric Mining: Scaling Geometric Hashing to Large Datasets

    Full text link
    It is known that relative feature location is important in representing objects, but assumptions that make learning tractable often simplify how structure is encoded e.g. spatial pooling or star models. For example, techniques such as spatial pyramid matching (SPM), in-conjunction with machine learning techniques perform well [13]. However, there are limitations to such spatial encoding schemes which discard important information about the layout of features. In contrast, we propose to use the object itself to choose the basis of the features in an object centric approach. In doing so we return to the early work of geometric hashing [18] but demonstrate how such approaches can be scaled-up to modern day object detection challenges in terms of both the number of examples and their variability. We apply a two stage process, initially filtering background features to localise the objects and then hashing the remaining pairwise features in an affine invariant model. During learning, we identify class-wise key feature predictors. We validate our detection and classification of objects on the PASCAL VOC'07 and' 11 [6] and CarDb [21] datasets and compare with state of the art detectors and classifiers. Importantly we demonstrate how structure in features can be efficiently identified and how its inclusion can increase performance. This feature centric learning technique allows us to localise objects even without object annotation during training and the resultant segmentation provides accurate state of the art object localization, without the need for annotations

    Interpretability and Explainability: A Machine Learning Zoo Mini-tour

    Full text link
    In this review, we examine the problem of designing interpretable and explainable machine learning models. Interpretability and explainability lie at the core of many machine learning and statistical applications in medicine, economics, law, and natural sciences. Although interpretability and explainability have escaped a clear universal definition, many techniques motivated by these properties have been developed over the recent 30 years with the focus currently shifting towards deep learning methods. In this review, we emphasise the divide between interpretability and explainability and illustrate these two different research directions with concrete examples of the state-of-the-art. The review is intended for a general machine learning audience with interest in exploring the problems of interpretation and explanation beyond logistic regression or random forest variable importance. This work is not an exhaustive literature survey, but rather a primer focusing selectively on certain lines of research which the authors found interesting or informative

    A Survey on Label-efficient Deep Image Segmentation: Bridging the Gap between Weak Supervision and Dense Prediction

    Full text link
    The rapid development of deep learning has made a great progress in image segmentation, one of the fundamental tasks of computer vision. However, the current segmentation algorithms mostly rely on the availability of pixel-level annotations, which are often expensive, tedious, and laborious. To alleviate this burden, the past years have witnessed an increasing attention in building label-efficient, deep-learning-based image segmentation algorithms. This paper offers a comprehensive review on label-efficient image segmentation methods. To this end, we first develop a taxonomy to organize these methods according to the supervision provided by different types of weak labels (including no supervision, inexact supervision, incomplete supervision and inaccurate supervision) and supplemented by the types of segmentation problems (including semantic segmentation, instance segmentation and panoptic segmentation). Next, we summarize the existing label-efficient image segmentation methods from a unified perspective that discusses an important question: how to bridge the gap between weak supervision and dense prediction -- the current methods are mostly based on heuristic priors, such as cross-pixel similarity, cross-label constraint, cross-view consistency, and cross-image relation. Finally, we share our opinions about the future research directions for label-efficient deep image segmentation.Comment: Accepted to IEEE TPAM

    Physics of three dimensional bosonic topological insulators: Surface Deconfined Criticality and Quantized Magnetoelectric Effect

    Full text link
    We discuss physical properties of `integer' topological phases of bosons in D=3+1 dimensions, protected by internal symmetries like time reversal and/or charge conservation. These phases invoke interactions in a fundamental way but do not possess topological order and are bosonic analogs of free fermion topological insulators and superconductors. While a formal cohomology based classification of such states was recently discovered, their physical properties remain mysterious. Here we develop a field theoretic description of several of these states and show that they possess unusual surface states, which if gapped, must either break the underlying symmetry, or develop topological order. In the latter case, symmetries are implemented in a way that is forbidden in a strictly two dimensional theory. While this is the usual fate of the surface states, exotic gapless states can also be realized. For example, tuning parameters can naturally lead to a deconfined quantum critical point or, in other situations, a fully symmetric vortex metal phase. We discuss cases where the topological phases are characterized by quantized magnetoelectric response \theta, which, somewhat surprisingly, is an odd multiple of 2\pi. Two different surface theories are shown to capture these phenomena - the first is a nonlinear sigma model with a topological term. The second invokes vortices on the surface that transform under a projective representation of the symmetry group. A bulk field theory consistent with these properties is identified, which is a multicomponent `BF' theory supplemented, crucially, with a topological term. A possible topological phase characterized by the thermal analog of the magnetoelectric effect is also discussed.Comment: 25 pages+ 3 pages Appendices, 3 figures. Introduction rewritten for clarity, minor technical changes and additional details of surface topological order adde

    Fine-Grained Linguistic Soft Constraints on Statistical Natural Language Processing Models

    Get PDF
    This dissertation focuses on effective combination of data-driven natural language processing (NLP) approaches with linguistic knowledge sources that are based on manual text annotation or word grouping according to semantic commonalities. I gainfully apply fine-grained linguistic soft constraints -- of syntactic or semantic nature -- on statistical NLP models, evaluated in end-to-end state-of-the-art statistical machine translation (SMT) systems. The introduction of semantic soft constraints involves intrinsic evaluation on word-pair similarity ranking tasks, extension from words to phrases, application in a novel distributional paraphrase generation technique, and an introduction of a generalized framework of which these soft semantic and syntactic constraints can be viewed as instances, and in which they can be potentially combined. Fine granularity is key in the successful combination of these soft constraints, in many cases. I show how to softly constrain SMT models by adding fine-grained weighted features, each preferring translation of only a specific syntactic constituent. Previous attempts using coarse-grained features yielded negative results. I also show how to softly constrain corpus-based semantic models of words (“distributional profiles”) to effectively create word-sense-aware models, by using semantic word grouping information found in a manually compiled thesaurus. Previous attempts, using hard constraints and resulting in aggregated, coarse-grained models, yielded lower gains. A novel paraphrase generation technique incorporating these soft semantic constraints is then also evaluated in a SMT system. This paraphrasing technique is based on the Distributional Hypothesis. The main advantage of this novel technique over current “pivoting” techniques for paraphrasing is the independence from parallel texts, which are a limited resource. The evaluation is done by augmenting translation models with paraphrase-based translation rules, where fine-grained scoring of paraphrase-based rules yields significantly higher gains. The model augmentation includes a novel semantic reinforcement component: In many cases there are alternative paths of generating a paraphrase-based translation rule. Each of these paths reinforces a dedicated score for the “goodness” of the new translation rule. This augmented score is then used as a soft constraint, in a weighted log-linear feature, letting the translation model learn how much to “trust” the paraphrase-based translation rules. The work reported here is the first to use distributional semantic similarity measures to improve performance of an end-to-end phrase-based SMT system. The unified framework for statistical NLP models with soft linguistic constraints enables, in principle, the combination of both semantic and syntactic constraints -- and potentially other constraints, too -- in a single SMT model

    Deep Neural Network Compression with Filter Pruning

    Get PDF
    The rapid development of convolutional neural networks (CNNs) in computer vision tasks has inspired researchers to apply their potential to embedded or mobile devices. However, it typically requires a large amount of computation and memory footprint, limiting their deployment in those resource-limited systems. Therefore, how to compress complex networks while maintaining competitive performance has become the focus of attention in recent years. On the subject of network compression, filter pruning methods that achieve structured compact model by finding and removing redundant filters, have attracted widespread attention. Inspired by previous dedicated works, this thesis focuses on the way to obtain the compact model while maximizing the retention of the original model performance. In particular, aiming at the limitations of choosing filters on the existing popular pruning methods, several novel filter pruning strategies are proposed to find and remove redundant filters more accurately to reduce the performance loss of the model caused by pruning. For instance, the filter pruning method with an attention mechanism (Chapter 3), data-dependent filter pruning guided by LSTM (Chapter 4), and filter pruning with uniqueness mechanism in the frequency domain (Chapter 5). This thesis first addresses the filter pruning issue from a global perspective. To this end, we propose a new scheme, termed Pruning Filter with an Attention Mechanism (PFAM). That is, by establishing the dependency/relationship between filters at each layer, we explore the long-term dependence between filters via attention module in order to choose the tobe-pruned filters. Unlike prior approaches that identify the to-be-pruned filters simply based on their intrinsic properties, the less correlated filters are first pruned after the pruning step in the current training epoch and then reconstructed and updated during the subsequent training epoch. Thus, the compressed network model can be achieved without the requirement for a pre-trained model since input data can be manipulated with the maximum information maintained when the original training strategy is executed. Next, it is noticed that most existing pruning algorithms seek to prune the filter layer by layer. Specifically, they guide filter pruning at each layer by setting a global pruning rate, which indicates that each convolutional layer is treated equally without regard to its depth and width. In this situation, we argue that the convolutional layers in the network also have varying degrees of significance. Besides, we propose that selecting the appropriate layers for pruning is more reasonable since it can result in more complexity reduction with less performance loss by keeping and removing more filters in those critical and nonsignificant layers, respectively. In order to do this, long short-term memory (LSTM) is employed to learn the hierarchical properties of a network and to generalize a global network pruning scheme. On top of that, we present a data-dependent soft pruning strategy named Squeeze-Excitation-Pruning (SEP), which does not physically prune any filters but removes specific kernels involved in calculating forward and backward propagations based on the pruning scheme. Doing so can further decrease the model’s performance decline while achieving a deep model compression. Lastly, we transfer the concept of relationship from the filter level to the feature map level because the feature maps can reflect the comprehensive information of both input data and filters. Hence, we propose Filter Pruning with Uniqueness Mechanism in the Frequency Domain (FPUM) to serve as a guideline for the filter pruning strategy by generating the correlation between feature maps. Specifically, we first transfer features to the frequency domain by Discrete Cosine Transform (DCT). Then, for each feature map, we compute a uniqueness score, which measures its probability of being replaced by others. Doing so allows us to prune the filters corresponding to the low-uniqueness maps without significant performance degradation. In addition, our strategy is more resistant to noise than spatial methods, further enhancing the network’s compactness while maintaining performance, as the critical pruning clues are more concentrated following DCT

    Leptonic Decays of Neutral B Mesons in the Three-Spurion Two-Higgs-Doublet Model

    Get PDF
    Leptonische Zerfälle Bq=d,sl+lB_{q=d,s} \to l^+ l^- neutraler BB-Mesonen sind im Standardmodell der Teilchenphysik stark unterdrückt und stellen daher einen idealen Prüfstand für Modelle neuer Physik dar. In dieser Arbeit wird ein das sogenannte Typ-II-Modell erweiterndes Zwei-Higgs-Dublett-Modell vorgestellt, welches im up-artigen Quark-Sektor flavourändernde Yukawa-Kopplungen der schweren neutralen Higgs-Bosonen beinhaltet. Dieses Modell wird als Drei-Spurion-Modell ohne diskrete Z2\mathbb{Z}_2-Symmetrie bezeichnet. Die zur Beschreibung mittels einer effektiven Feldtheorie benötigten Wilson-Koeffizienten für Bql+lB_q \to l^+ l^- werden in dieser Arbeit zur nächstführenden Zwei-Schleifen-Ordnung in QCD berechnet. Besonderes Augenmerk wird auf die Renormierung des vorgestellten Modells im Kontext dieser Zerfällt gelegt, insbesondere auf die konzeptionellen Unterschiede zu den besser bekannten Z2\mathbb{Z}_2-symmetrischen Modellen. Weiter werden relevante Einschränkungen aus der Flavour-Physik und von Beschleuniger-Experimenten präsentiert und daraus Informationen über den erlaubten Parameterbereich der am stärksten eingehenden Kopplungen abgeleitet. Insbesondere wird in dieser Dissertation gezeigt, dass selbst sehr kleine Werte mancher dieser Kopplungen beträchtlichen Einfluss auf Observablen wie das Zerfallsverhältnis in Bsμ+μB_s \to \mu^+ \mu^- haben können
    corecore