21,506 research outputs found

    New efficient algorithms for multiple change-point detection with kernels

    Get PDF
    Several statistical approaches based on reproducing kernels have been proposed to detect abrupt changes arising in the full distribution of the observations and not only in the mean or variance. Some of these approaches enjoy good statistical properties (oracle inequality, \ldots). Nonetheless, they have a high computational cost both in terms of time and memory. This makes their application difficult even for small and medium sample sizes (n<104n< 10^4). This computational issue is addressed by first describing a new efficient and exact algorithm for kernel multiple change-point detection with an improved worst-case complexity that is quadratic in time and linear in space. It allows dealing with medium size signals (up to n105n \approx 10^5). Second, a faster but approximation algorithm is described. It is based on a low-rank approximation to the Gram matrix. It is linear in time and space. This approximation algorithm can be applied to large-scale signals (n106n \geq 10^6). These exact and approximation algorithms have been implemented in \texttt{R} and \texttt{C} for various kernels. The computational and statistical performances of these new algorithms have been assessed through empirical experiments. The runtime of the new algorithms is observed to be faster than that of other considered procedures. Finally, simulations confirmed the higher statistical accuracy of kernel-based approaches to detect changes that are not only in the mean. These simulations also illustrate the flexibility of kernel-based approaches to analyze complex biological profiles made of DNA copy number and allele B frequencies. An R package implementing the approach will be made available on github

    Trainable COSFIRE filters for vessel delineation with application to retinal images

    Get PDF
    Retinal imaging provides a non-invasive opportunity for the diagnosis of several medical pathologies. The automatic segmentation of the vessel tree is an important pre-processing step which facilitates subsequent automatic processes that contribute to such diagnosis. We introduce a novel method for the automatic segmentation of vessel trees in retinal fundus images. We propose a filter that selectively responds to vessels and that we call B-COSFIRE with B standing for bar which is an abstraction for a vessel. It is based on the existing COSFIRE (Combination Of Shifted Filter Responses) approach. A B-COSFIRE filter achieves orientation selectivity by computing the weighted geometric mean of the output of a pool of Difference-of-Gaussians filters, whose supports are aligned in a collinear manner. It achieves rotation invariance efficiently by simple shifting operations. The proposed filter is versatile as its selectivity is determined from any given vessel-like prototype pattern in an automatic configuration process. We configure two B-COSFIRE filters, namely symmetric and asymmetric, that are selective for bars and bar-endings, respectively. We achieve vessel segmentation by summing up the responses of the two rotation-invariant B-COSFIRE filters followed by thresholding. The results that we achieve on three publicly available data sets (DRIVE: Se = 0.7655, Sp = 0.9704; STARE: Se = 0.7716, Sp = 0.9701; CHASE_DB1: Se = 0.7585, Sp = 0.9587) are higher than many of the state-of-the-art methods. The proposed segmentation approach is also very efficient with a time complexity that is significantly lower than existing methods.peer-reviewe

    T-PHOT: A new code for PSF-matched, prior-based, multiwavelength extragalactic deconfusion photometry

    Get PDF
    We present T-PHOT, a publicly available software aimed at extracting accurate photometry from low-resolution images of deep extragalactic fields, where the blending of sources can be a serious problem for the accurate and unbiased measurement of fluxes and colours. T-PHOT has been developed within the ASTRODEEP project and it can be considered as the next generation to TFIT, providing significant improvements above it and other similar codes. T-PHOT gathers data from a high-resolution image of a region of the sky, and uses it to obtain priors for the photometric analysis of a lower resolution image of the same field. It can handle different types of datasets as input priors: i) a list of objects that will be used to obtain cutouts from the real high-resolution image; ii) a set of analytical models; iii) a list of unresolved, point-like sources, useful e.g. for far-infrared wavelength domains. We show that T-PHOT yields accurate estimations of fluxes within the intrinsic uncertainties of the method, when systematic errors are taken into account (which can be done thanks to a flagging code given in the output). T-PHOT is many times faster than similar codes like TFIT and CONVPHOT (up to hundreds, depending on the problem and the method adopted), whilst at the same time being more robust and more versatile. This makes it an optimal choice for the analysis of large datasets. In addition we show how the use of different settings and methods significantly enhances the performance. Given its versatility and robustness, T-PHOT can be considered the preferred choice for combined photometric analysis of current and forthcoming extragalactic optical to far-infrared imaging surveys. [abridged]Comment: 23 pages, 20 figures, 2 table

    Image Segmentation Using Active Contours Driven by the Bhattacharyya Gradient Flow

    Get PDF
    ©2007 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or distribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE. This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.DOI: 10.1109/TIP.2007.908073This paper addresses the problem of image segmentation by means of active contours, whose evolution is driven by the gradient flow derived froman energy functional that is based on the Bhattacharyya distance. In particular, given the values of a photometric variable (or of a set thereof), which is to be used for classifying the image pixels, the active contours are designed to converge to the shape that results in maximal discrepancy between the empirical distributions of the photometric variable inside and outside of the contours. The above discrepancy is measured by means of the Bhattacharyya distance that proves to be an extremely useful tool for solving the problem at hand. The proposed methodology can be viewed as a generalization of the segmentation methods, in which active contours maximize the difference between a finite number of empirical moments of the "inside" and "outside" distributions. Furthermore, it is shown that the proposed methodology is very versatile and flexible in the sense that it allows one to easily accommodate a diversity of the image features based on which the segmentation should be performed. As an additional contribution, a method for automatically adjusting the smoothness properties of the empirical distributions is proposed. Such a procedure is crucial in situations when the number of data samples (supporting a certain segmentation class) varies considerably in the course of the evolution of the active contour. In this case, the smoothness properties of the empirical distributions have to be properly adjusted to avoid either over- or underestimation artifacts. Finally, a number of relevant segmentation results are demonstrated and some further research directions are discussed

    Cross-Talk-Free Multi-Color STORM Imaging Using a Single Fluorophore

    Get PDF
    Multi-color stochastic optical reconstruction microscopy (STORM) is routinely performed; however, the various approaches for achieving multiple colors have important caveats. Color cross-talk, limited availability of spectrally distinct fluorophores with optimal brightness and duty cycle, incompatibility of imaging buffers for different fluorophores, and chromatic aberrations impact the spatial resolution and ultimately the number of colors that can be achieved. We overcome these complexities and develop a simple approach for multi-color STORM imaging using a single fluorophore and sequential labelling. In addition, we present a simple and versatile method to locate the same region of interest on different days and even on different microscopes. In combination, these approaches enable cross-talk-free multi-color imaging of sub-cellular structures.Peer ReviewedPostprint (published version

    Ultrasound IMT measurement on a multi-ethnic and multi-institutional database: Our review and experience using four fully automated and one semi-automated methods

    Get PDF
    Automated and high performance carotid intima-media thickness (IMT) measurement is gaining increasing importance in clinical practice to assess the cardiovascular risk of patients. In this paper, we compare four fully automated IMT measurement techniques (CALEX, CAMES, CARES and CAUDLES) and one semi-automated technique (FOAM). We present our experience using these algorithms, whose lumen-intima and media-adventitia border estimation use different methods that can be: (a) edge-based; (b) training-based; (c) feature-based; or (d) directional Edge-Flow based. Our database (DB) consisted of 665 images that represented a multi-ethnic group and was acquired using four OEM scanners. The performance evaluation protocol adopted error measures, reproducibility measures, and Figure of Merit (FoM). FOAM showed the best performance, with an IMT bias equal to 0.025 ± 0.225 mm, and a FoM equal to 96.6%. Among the four automated methods, CARES showed the best results with a bias of 0.032 ± 0.279 mm, and a FoM to 95.6%, which was statistically comparable to that of FOAM performance in terms of accuracy and reproducibility. This is the first time that completely automated and user-driven techniques have been compared on a multi-ethnic dataset, acquired using multiple original equipment manufacturer (OEM) machines with different gain settings, representing normal and pathologic case

    Point Pair Feature based Object Detection for Random Bin Picking

    Full text link
    Point pair features are a popular representation for free form 3D object detection and pose estimation. In this paper, their performance in an industrial random bin picking context is investigated. A new method to generate representative synthetic datasets is proposed. This allows to investigate the influence of a high degree of clutter and the presence of self similar features, which are typical to our application. We provide an overview of solutions proposed in literature and discuss their strengths and weaknesses. A simple heuristic method to drastically reduce the computational complexity is introduced, which results in improved robustness, speed and accuracy compared to the naive approach
    corecore