44 research outputs found

    Crowd counting and segmentation in visual surveillance

    Get PDF
    Reference no. MP-PD.8In this paper, the crowd counting and segmentation problem is formulated as a maximum a posterior problem, in which 3D human shape models are designed and matched with image evidence provided by foreground/background separation and probability of boundary. The solution is obtained by considering only the human candidates that are possible to be un-occluded in each iteration, and then applying on them a validation and rejection strategy based on minimum description length. The merit of the proposed optimization procedure is that its computational cost is much smaller than that of the global optimization methods while its performance is comparable to them. The approach is shown to be robust with respect to severe partial occlusions. ©2009 IEEE.published_or_final_versionThe 16th IEEE International Conference on Image Processing (ICIP 2009), Cairo, Egypt, 7-10 November 2009. In International Conference on Image Processing Proceedings, 2009, p. 2573-257

    Quantification of structures of skin lesion and calibration of dermoscopy images for automated melanoma diagnosis

    Get PDF
    研究成果の概要 (和文) : 我々が開発している、インターネット上の悪性黒色腫自動診断支援システムの実用化を具体的に推進するために、課題名にある2点の開発を重点的に行った。(1)定量化研究においては、皮膚科医が臨床現場で用いるABCD rule、7-point checklistの全15項目について、皮膚科医の判断と統計的有意差が見られないモデルの構築に成功した。(2)画像補正研究においては、特殊なハードウェアなどを必要とせず、画像の明度、色彩を適切なダーモスコピー画像のものに調整する手法を確立した。研究成果の概要 (英文) : I had investigated mainly following two themes for making our Internet-based melanoma screening system fit for practical use. (1) Quantification of dermoscopic structures (2) Development of automated color calibration for dermoscopy images. For both themes, I had successfully achieved the objectives set out in the proposal. (1) Succeeded to build a recognition system that has statistically no difference with expert dermatologists in recognizing a total of 15 dermoscopic structures defined in ABCD rule and 7-point checklist. (2) Achieved equivalent color calibration performance for dermoscopy images in luminance, hue and saturation without any special devices

    Direction-adaptive grey-level morphology. Application to 3D vascular brain imaging

    Get PDF
    International audienceSegmentation and analysis of blood vessels is an important issue in medical imaging. In 3D cerebral angiographic data, the vascular signal is however hard to accurately detect and can, in particular, be disconnected. In this article, we present a procedure utilising both linear, Hessian-based and morphological methods for blood vessel edge enhancement and reconnection. More specifically, multi-scale second-order derivative analysis is performed to detect candidate vessels as well as their orientation. This information is then fed to a spatially-variant morphological filter for reconnection and reconstruction. The result is a fast and effective vessel-reconnecting method

    Multiplicative Noise Removal Using Variable Splitting and Constrained Optimization

    Full text link
    Multiplicative noise (also known as speckle noise) models are central to the study of coherent imaging systems, such as synthetic aperture radar and sonar, and ultrasound and laser imaging. These models introduce two additional layers of difficulties with respect to the standard Gaussian additive noise scenario: (1) the noise is multiplied by (rather than added to) the original image; (2) the noise is not Gaussian, with Rayleigh and Gamma being commonly used densities. These two features of multiplicative noise models preclude the direct application of most state-of-the-art algorithms, which are designed for solving unconstrained optimization problems where the objective has two terms: a quadratic data term (log-likelihood), reflecting the additive and Gaussian nature of the noise, plus a convex (possibly nonsmooth) regularizer (e.g., a total variation or wavelet-based regularizer/prior). In this paper, we address these difficulties by: (1) converting the multiplicative model into an additive one by taking logarithms, as proposed by some other authors; (2) using variable splitting to obtain an equivalent constrained problem; and (3) dealing with this optimization problem using the augmented Lagrangian framework. A set of experiments shows that the proposed method, which we name MIDAL (multiplicative image denoising by augmented Lagrangian), yields state-of-the-art results both in terms of speed and denoising performance.Comment: 11 pages, 7 figures, 2 tables. To appear in the IEEE Transactions on Image Processing

    Recognition and Classification of Fast Food Images

    Get PDF
    Image processing is widely used for food recognition. A lot of different algorithms regarding food identification and classification have been proposed in recent research works. In this paper, we have use an easy and one of the most powerful machine learning technique from the field of deep learning to recognize and classify different categories of fast food images. We have used a pre trained Convolutional Neural Network (CNN) as a feature extractor to train an image category classifier. CNN2019;s can learn rich feature representations which often perform much better than other handcrafted features such as histogram of oriented gradients (HOG), Local binary patterns (LBP), or speeded up robust features (SURF). A multiclass linear Support Vector Machine (SVM) classifier trained with extracted CNN features is used to classify fast food images to ten different classes. After working on two different benchmark databases, we got the success rate of 99.5% which is higher than the accuracy achieved using bag of features (BoF) and SURF

    Consistent Image Decoding from Multiple Lossy Versions

    Get PDF
    With the recent development of tools for data sharing in social networks and peer to peer networks, the same information is often stored in different nodes. Peer-to-peer protocols usually allow one user to collect portions of the same file from different nodes in the network, substantially improving the rate at which data are received by the end user. In some cases, however, the same multimedia document is available in different lossy versions on the network nodes. In such situations, one may be interested in collecting all available versions of the same document and jointly decoding them to obtain a better reconstruction of the original. In this paper we study some methods to jointly decode different versions of the same image. We compare different uses of the method of Projections Onto Convex Sets (POCS) with some Convex Optimization techniques in order to reconstruct an image for which JPEG and JPEG2000 lossy versions are available

    On the use of compressed sensing techniques for improving multistatic millimeter-wave portal-based personnel screening

    Get PDF
    This work develops compressed sensing techniques to improve the performance of an active three dimensional (3D) millimeter wave imaging system for personnel security screening. The system is able to produce a high-resolution 3D reconstruction of the whole human body surface and reveal concealed objects under clothing. Innovative multistatic millimeter wave radar designs and algorithms, which have been previously validated, are combined to improve the reconstruction results over previous approaches. Compressed Sensing techniques are used to drastically reduce the number of sensors, thus simplifying the system design and fabrication. Representative simulation results showing good performance of the proposed system are provided and supported by several sample measurement

    Super-resolution in map-making based on a physical instrument model and regularized inversion. Application to SPIRE/Herschel

    Full text link
    We investigate super-resolution methods for image reconstruction from data provided by a family of scanning instruments like the Herschel observatory. To do this, we constructed a model of the instrument that faithfully reflects the physical reality, accurately taking the acquisition process into account to explain the data in a reliable manner. The inversion, ie the image reconstruction process, is based on a linear approach resulting from a quadratic regularized criterion and numerical optimization tools. The application concerns the reconstruction of maps for the SPIRE instrument of the Herschel observatory. The numerical evaluation uses simulated and real data to compare the standard tool (coaddition) and the proposed method. The inversion approach is capable to restore spatial frequencies over a bandwidth four times that possible with coaddition and thus to correctly show details invisible on standard maps. The approach is also applied to real data with significant improvement in spatial resolution.Comment: Astronomy & Astrophysic

    An Efficient Direction Field-Based Method for the Detection of Fasteners on High-Speed Railways

    Get PDF
    Railway inspection is an important task in railway maintenance to ensure safety. The fastener is a major part of the railway which fastens the tracks to the ground. The current article presents an efficient method to detect fasteners on the basis of image processing and pattern recognition techniques, which can be used to detect the absence of fasteners on the corresponding track in high-speed(up to 400 km/h). The Direction Field is extracted as the feature descriptor for recognition. In addition, the appropriate weight coefficient matrix is presented for robust and rapid matching in a complex environment. Experimental results are presented to show that the proposed method is computation efficient and robust for the detection of fasteners in a complex environment. Through the practical device fixed on the track inspection train, enough fastener samples are obtained, and the feasibility of the method is verified at 400 km/h
    corecore