1,607 research outputs found

    A variational algorithm for the detection of line segments

    Get PDF
    In this paper we propose an algorithm for the detection of edges in images that is based on topological asymptotic analysis. Motivated from the Mumford--Shah functional, we consider a variational functional that penalizes oscillations outside some approximate edge set, which we represent as the union of a finite number of thin strips, the width of which is an order of magnitude smaller than their length. In order to find a near optimal placement of these strips, we compute an asymptotic expansion of the functional with respect to the strip size. This expansion is then employed for defining a (topological) gradient descent like minimization method. As opposed to a recently proposed method by some of the authors, which uses coverings with balls, the usage of strips includes some directional information into the method, which can be used for obtaining finer edges and can also result in a reduction of computation times

    Discretization of variational regularization in Banach spaces

    Full text link
    Consider a nonlinear ill-posed operator equation F(u)=yF(u)=y where FF is defined on a Banach space XX. In general, for solving this equation numerically, a finite dimensional approximation of XX and an approximation of FF are required. Moreover, in general the given data \yd of yy are noisy. In this paper we analyze finite dimensional variational regularization, which takes into account operator approximations and noisy data: We show (semi-)convergence of the regularized solution of the finite dimensional problems and establish convergence rates in terms of Bregman distances under appropriate sourcewise representation of a solution of the equation. The more involved case of regularization in nonseparable Banach spaces is discussed in detail. In particular we consider the space of finite total variation functions, the space of functions of finite bounded deformation, and the LL^\infty--space

    Necessary conditions for variational regularization schemes

    Full text link
    We study variational regularization methods in a general framework, more precisely those methods that use a discrepancy and a regularization functional. While several sets of sufficient conditions are known to obtain a regularization method, we start with an investigation of the converse question: How could necessary conditions for a variational method to provide a regularization method look like? To this end, we formalize the notion of a variational scheme and start with comparison of three different instances of variational methods. Then we focus on the data space model and investigate the role and interplay of the topological structure, the convergence notion and the discrepancy functional. Especially, we deduce necessary conditions for the discrepancy functional to fulfill usual continuity assumptions. The results are applied to discrepancy functionals given by Bregman distances and especially to the Kullback-Leibler divergence.Comment: To appear in Inverse Problem

    Parkinson's disease biomarkers: perspective from the NINDS Parkinson's Disease Biomarkers Program

    Get PDF
    Biomarkers for Parkinson's disease (PD) diagnosis, prognostication and clinical trial cohort selection are an urgent need. While many promising markers have been discovered through the National Institute of Neurological Disorders and Stroke Parkinson's Disease Biomarker Program (PDBP) and other mechanisms, no single PD marker or set of markers are ready for clinical use. Here we discuss the current state of biomarker discovery for platforms relevant to PDBP. We discuss the role of the PDBP in PD biomarker identification and present guidelines to facilitate their development. These guidelines include: harmonizing procedures for biofluid acquisition and clinical assessments, replication of the most promising biomarkers, support and encouragement of publications that report negative findings, longitudinal follow-up of current cohorts including the PDBP, testing of wearable technologies to capture readouts between study visits and development of recently diagnosed (de novo) cohorts to foster identification of the earliest markers of disease onset

    Iteratively regularized Newton-type methods for general data misfit functionals and applications to Poisson data

    Get PDF
    We study Newton type methods for inverse problems described by nonlinear operator equations F(u)=gF(u)=g in Banach spaces where the Newton equations F(un;un+1un)=gF(un)F'(u_n;u_{n+1}-u_n) = g-F(u_n) are regularized variationally using a general data misfit functional and a convex regularization term. This generalizes the well-known iteratively regularized Gauss-Newton method (IRGNM). We prove convergence and convergence rates as the noise level tends to 0 both for an a priori stopping rule and for a Lepski{\u\i}-type a posteriori stopping rule. Our analysis includes previous order optimal convergence rate results for the IRGNM as special cases. The main focus of this paper is on inverse problems with Poisson data where the natural data misfit functional is given by the Kullback-Leibler divergence. Two examples of such problems are discussed in detail: an inverse obstacle scattering problem with amplitude data of the far-field pattern and a phase retrieval problem. The performence of the proposed method for these problems is illustrated in numerical examples

    Measurement of the top quark-pair production cross section with ATLAS in pp collisions at \sqrt{s}=7\TeV

    Get PDF
    A measurement of the production cross-section for top quark pairs(\ttbar) in pppp collisions at \sqrt{s}=7 \TeV is presented using data recorded with the ATLAS detector at the Large Hadron Collider. Events are selected in two different topologies: single lepton (electron ee or muon μ\mu) with large missing transverse energy and at least four jets, and dilepton (eeee, μμ\mu\mu or eμe\mu) with large missing transverse energy and at least two jets. In a data sample of 2.9 pb-1, 37 candidate events are observed in the single-lepton topology and 9 events in the dilepton topology. The corresponding expected backgrounds from non-\ttbar Standard Model processes are estimated using data-driven methods and determined to be 12.2±3.912.2 \pm 3.9 events and 2.5±0.62.5 \pm 0.6 events, respectively. The kinematic properties of the selected events are consistent with SM \ttbar production. The inclusive top quark pair production cross-section is measured to be \sigmattbar=145 \pm 31 ^{+42}_{-27} pb where the first uncertainty is statistical and the second systematic. The measurement agrees with perturbative QCD calculations.Comment: 30 pages plus author list (50 pages total), 9 figures, 11 tables, CERN-PH number and final journal adde

    Inclusive search for same-sign dilepton signatures in pp collisions at root s=7 TeV with the ATLAS detector

    Get PDF
    An inclusive search is presented for new physics in events with two isolated leptons (e or mu) having the same electric charge. The data are selected from events collected from p p collisions at root s = 7 TeV by the ATLAS detector and correspond to an integrated luminosity of 34 pb(-1). The spectra in dilepton invariant mass, missing transverse momentum and jet multiplicity are presented and compared to Standard Model predictions. In this event sample, no evidence is found for contributions beyond those of the Standard Model. Limits are set on the cross-section in a fiducial region for new sources of same-sign high-mass dilepton events in the ee, e mu and mu mu channels. Four models predicting same-sign dilepton signals are constrained: two descriptions of Majorana neutrinos, a cascade topology similar to supersymmetry or universal extra dimensions, and fourth generation d-type quarks. Assuming a new physics scale of 1 TeV, Majorana neutrinos produced by an effective operator V with masses below 460 GeV are excluded at 95% confidence level. A lower limit of 290 GeV is set at 95% confidence level on the mass of fourth generation d-type quarks

    Measurement of the production of a W boson in association with a charm quark in pp collisions at √s = 7 TeV with the ATLAS detector

    Get PDF
    The production of a W boson in association with a single charm quark is studied using 4.6 fb−1 of pp collision data at s√ = 7 TeV collected with the ATLAS detector at the Large Hadron Collider. In events in which a W boson decays to an electron or muon, the charm quark is tagged either by its semileptonic decay to a muon or by the presence of a charmed meson. The integrated and differential cross sections as a function of the pseudorapidity of the lepton from the W-boson decay are measured. Results are compared to the predictions of next-to-leading-order QCD calculations obtained from various parton distribution function parameterisations. The ratio of the strange-to-down sea-quark distributions is determined to be 0.96+0.26−0.30 at Q 2 = 1.9 GeV2, which supports the hypothesis of an SU(3)-symmetric composition of the light-quark sea. Additionally, the cross-section ratio σ(W + +c¯¯)/σ(W − + c) is compared to the predictions obtained using parton distribution function parameterisations with different assumptions about the s−s¯¯¯ quark asymmetry

    Low Complexity Regularization of Linear Inverse Problems

    Full text link
    Inverse problems and regularization theory is a central theme in contemporary signal processing, where the goal is to reconstruct an unknown signal from partial indirect, and possibly noisy, measurements of it. A now standard method for recovering the unknown signal is to solve a convex optimization problem that enforces some prior knowledge about its structure. This has proved efficient in many problems routinely encountered in imaging sciences, statistics and machine learning. This chapter delivers a review of recent advances in the field where the regularization prior promotes solutions conforming to some notion of simplicity/low-complexity. These priors encompass as popular examples sparsity and group sparsity (to capture the compressibility of natural signals and images), total variation and analysis sparsity (to promote piecewise regularity), and low-rank (as natural extension of sparsity to matrix-valued data). Our aim is to provide a unified treatment of all these regularizations under a single umbrella, namely the theory of partial smoothness. This framework is very general and accommodates all low-complexity regularizers just mentioned, as well as many others. Partial smoothness turns out to be the canonical way to encode low-dimensional models that can be linear spaces or more general smooth manifolds. This review is intended to serve as a one stop shop toward the understanding of the theoretical properties of the so-regularized solutions. It covers a large spectrum including: (i) recovery guarantees and stability to noise, both in terms of 2\ell^2-stability and model (manifold) identification; (ii) sensitivity analysis to perturbations of the parameters involved (in particular the observations), with applications to unbiased risk estimation ; (iii) convergence properties of the forward-backward proximal splitting scheme, that is particularly well suited to solve the corresponding large-scale regularized optimization problem

    Search for direct production of charginos and neutralinos in events with three leptons and missing transverse momentum in √s = 7 TeV pp collisions with the ATLAS detector

    Get PDF
    A search for the direct production of charginos and neutralinos in final states with three electrons or muons and missing transverse momentum is presented. The analysis is based on 4.7 fb−1 of proton–proton collision data delivered by the Large Hadron Collider and recorded with the ATLAS detector. Observations are consistent with Standard Model expectations in three signal regions that are either depleted or enriched in Z-boson decays. Upper limits at 95% confidence level are set in R-parity conserving phenomenological minimal supersymmetric models and in simplified models, significantly extending previous results
    corecore