973 research outputs found
Vanishing Point Detection with Direct and Transposed Fast Hough Transform inside the neural network
In this paper, we suggest a new neural network architecture for vanishing
point detection in images. The key element is the use of the direct and
transposed Fast Hough Transforms separated by convolutional layer blocks with
standard activation functions. It allows us to get the answer in the
coordinates of the input image at the output of the network and thus to
calculate the coordinates of the vanishing point by simply selecting the
maximum. Besides, it was proved that calculation of the transposed Fast Hough
Transform can be performed using the direct one. The use of integral operators
enables the neural network to rely on global rectilinear features in the image,
and so it is ideal for detecting vanishing points. To demonstrate the
effectiveness of the proposed architecture, we use a set of images from a DVR
and show its superiority over existing methods. Note, in addition, that the
proposed neural network architecture essentially repeats the process of direct
and back projection used, for example, in computed tomography.Comment: 9 pages, 9 figures, submitted to "Computer Optics"; extra experiment
added, new theorem proof added, references added; typos correcte
Simulation Models for Straight Lines Images Detection Using Hough Transform
The Hough transform (HT) is a robust parameter estimator of multidimensional features in images. The HT is an established technique which evidences a shape by mapping image edge points into a parameter space. Recently, the
formulation of the HT has been extended to extract analytic arbitrary shapes which change their appearance according to similarity transformations. It finds many applications in astronomical data analysis. It enables, in particular, to develop autoadaptive, fast algorithms for the detection of automated arc line identification. The HT is a technique which is used to isolate curves of a given shape in an image. The classical HT requires that the curve be specified in some parametric form and, hence is most commonly used in the detection of regular curves. The HT has been generalized so that it is capable of detecting arbitrary curved shapes
Fast Mojette Transform for Discrete Tomography
A new algorithm for reconstructing a two dimensional object from a set of one
dimensional projected views is presented that is both computationally exact and
experimentally practical. The algorithm has a computational complexity of O(n
log2 n) with n = N^2 for an NxN image, is robust in the presence of noise and
produces no artefacts in the reconstruction process, as is the case with
conventional tomographic methods. The reconstruction process is approximation
free because the object is assumed to be discrete and utilizes fully discrete
Radon transforms. Noise in the projection data can be suppressed further by
introducing redundancy in the reconstruction. The number of projections
required for exact reconstruction and the response to noise can be controlled
without comprising the digital nature of the algorithm. The digital projections
are those of the Mojette Transform, a form of discrete linogram. A simple
analytical mapping is developed that compacts these projections exactly into
symmetric periodic slices within the Discrete Fourier Transform. A new digital
angle set is constructed that allows the periodic slices to completely fill all
of the objects Discrete Fourier space. Techniques are proposed to acquire these
digital projections experimentally to enable fast and robust two dimensional
reconstructions.Comment: 22 pages, 13 figures, Submitted to Elsevier Signal Processin
Time-frequency transforms of white noises and Gaussian analytic functions
A family of Gaussian analytic functions (GAFs) has recently been linked to
the Gabor transform of white Gaussian noise [Bardenet et al., 2017]. This
answered pioneering work by Flandrin [2015], who observed that the zeros of the
Gabor transform of white noise had a very regular distribution and proposed
filtering algorithms based on the zeros of a spectrogram. The mathematical link
with GAFs provides a wealth of probabilistic results to inform the design of
such signal processing procedures. In this paper, we study in a systematic way
the link between GAFs and a class of time-frequency transforms of Gaussian
white noises on Hilbert spaces of signals. Our main observation is a conceptual
correspondence between pairs (transform, GAF) and generating functions for
classical orthogonal polynomials. This correspondence covers some classical
time-frequency transforms, such as the Gabor transform and the Daubechies-Paul
analytic wavelet transform. It also unveils new windowed discrete Fourier
transforms, which map white noises to fundamental GAFs. All these transforms
may thus be of interest to the research program `filtering with zeros'. We also
identify the GAF whose zeros are the extrema of the Gabor transform of the
white noise and derive their first intensity. Moreover, we discuss important
subtleties in defining a white noise and its transform on infinite dimensional
Hilbert spaces. Finally, we provide quantitative estimates concerning the
finite-dimensional approximations of these white noises, which is of practical
interest when it comes to implementing signal processing algorithms based on
GAFs.Comment: to appear in Applied and Computational Harmonic Analysi
Recommended from our members
The treatment of corner and pole-type singularities in numerical conformal mapping techniques
This paper is a report of recent developments concerning the nature and the treatment of singularities that affect certain numerical conformal mapping techniques. The paper also includes some new results on the nature of singularities that the mapping function may have in the complement of the closure of the domain under consideration
HoughNet: neural network architecture for vanishing points detection
In this paper we introduce a novel neural network architecture based on Fast
Hough Transform layer. The layer of this type allows our neural network to
accumulate features from linear areas across the entire image instead of local
areas. We demonstrate its potential by solving the problem of vanishing points
detection in the images of documents. Such problem occurs when dealing with
camera shots of the documents in uncontrolled conditions. In this case, the
document image can suffer several specific distortions including projective
transform. To train our model, we use MIDV-500 dataset and provide testing
results. The strong generalization ability of the suggested method is proven
with its applying to a completely different ICDAR 2011 dewarping contest. In
previously published papers considering these dataset authors measured the
quality of vanishing point detection by counting correctly recognized words
with open OCR engine Tesseract. To compare with them, we reproduce this
experiment and show that our method outperforms the state-of-the-art result.Comment: 6 pages, 6 figures, 2 tables, 28 references, conferenc
Robust approach to object recognition through fuzzy clustering and hough transform based methods
Object detection from two dimensional intensity images as well as three dimensional range images is considered. The emphasis is on the robust detection of shapes such as cylinders, spheres, cones, and planar surfaces, typically found in mechanical and manufacturing engineering applications. Based on the analyses of different HT methods, a novel method, called the Fast Randomized Hough Transform (FRHT) is proposed. The key idea of FRHT is to divide the original image into multiple regions and apply random sampling method to map data points in the image space into the parameter space or feature space, then obtain the parameters of true clusters. This results in the following characteristics, which are highly desirable in any method: high computation speed, low memory requirement, high result resolution and infinite parameter space. This project also considers use of fuzzy clustering techniques, such as Fuzzy C Quadric Shells (FCQS) clustering algorithm but combines the concept of noise prototype to form the Noise FCQS clustering algorithm that is robust against noise. Then a novel integrated clustering algorithm combining the advantages of FRHT and NFCQS methods is proposed. It is shown to be a robust clustering algorithm having the distinct advantages such as: the number of clusters need not be known in advance, the results are initialization independent, the detection accuracy is greatly improved, and the computation speed is very fast. Recent concepts from robust statistics, such as least trimmed squares estimation (LTS), minimum volume ellipsoid estimator (MVE) and the generalized MVE are also utilized to form a new robust algorithm called the generalized LTS for Quadric Surfaces (GLTS-QS) algorithm is developed. The experimental results indicate that the clustering method combining the FRHT and the GLTS-QS can improve clustering performance. Moreover, a new cluster validity method for circular clusters is proposed by considering the distribution of the points on the circular edge. Different methods for the computation of distance of a point from a cluster boundary, a common issue in all the range image clustering algorithms, are also discussed. The performance of all these algorithms is tested using various real and synthetic range and intensity images. The application of the robust clustering methods to the experimental granular flow research is also included
Seeing things
This paper is concerned with the problem of attaching meaningful symbols to aspects of the visible environment in machine and biological vision. It begins with a review of some of the arguments commonly used to support either the 'symbolic' or the 'behaviourist' approach to vision. Having explored these avenues without arriving at a satisfactory conclusion, we then present a novel argument, which starts from the question : given a functional description of a vision system, when could it be said to support a symbolic interpretation? We argue that to attach symbols to a system, its behaviour must exhibit certain well defined regularities in its response to its visual input and these are best described in terms of invariance and equivariance to transformations which act in the world and induce corresponding changes of the vision system state. This approach is illustrated with a brief exploration of the problem of identifying and acquiring visual representations having these symmetry properties, which also highlights the advantages of using an 'active' model of vision
- …