8 research outputs found

    Laplacian-Steered Neural Style Transfer

    Full text link
    Neural Style Transfer based on Convolutional Neural Networks (CNN) aims to synthesize a new image that retains the high-level structure of a content image, rendered in the low-level texture of a style image. This is achieved by constraining the new image to have high-level CNN features similar to the content image, and lower-level CNN features similar to the style image. However in the traditional optimization objective, low-level features of the content image are absent, and the low-level features of the style image dominate the low-level detail structures of the new image. Hence in the synthesized image, many details of the content image are lost, and a lot of inconsistent and unpleasing artifacts appear. As a remedy, we propose to steer image synthesis with a novel loss function: the Laplacian loss. The Laplacian matrix ("Laplacian" in short), produced by a Laplacian operator, is widely used in computer vision to detect edges and contours. The Laplacian loss measures the difference of the Laplacians, and correspondingly the difference of the detail structures, between the content image and a new image. It is flexible and compatible with the traditional style transfer constraints. By incorporating the Laplacian loss, we obtain a new optimization objective for neural style transfer named Lapstyle. Minimizing this objective will produce a stylized image that better preserves the detail structures of the content image and eliminates the artifacts. Experiments show that Lapstyle produces more appealing stylized images with less artifacts, without compromising their "stylishness".Comment: Accepted by the ACM Multimedia Conference (MM) 2017. 9 pages, 65 figure

    Segmentation of color images based on the gravitational clustering concept

    Get PDF
    A new clustering algorithm derived from the Markovian model of the gravitational clustering concept is proposed that works in the RGB measurement space for color image. To enable the model to be applicable in image segmentation, the new algorithm imposes a clustering constraint at each clustering iteration to control and determine the formation of multiple clusters. Using such constraint to limit the attraction between clusters, a termination condition can be easily defined. The new clustering algorithm is evaluated objectively and subjectively on three different images against the K-means clustering algorithm, the recursive histogram clustering algorithm for color (also known as the multi-spectral thresholding), the Hedley-Yan algorithm, and the widely used seed-based region growing algorithm. From the evaluation, it is observed that the new algorithm exhibits the following characteristics: (1) its objective measurement figures are comparable with the best in this group of segmentation algorithms; (2) it generates smoother region boundaries; (3) the segmented boundaries align closely with the original boundaries; and (4) it forms a meaningful number of segmented regions. © 1998 Society of Photo-Optical Instrumentation Engineers.published_or_final_versio

    Dynamic Block-Based Parameter Estimation for MRF Classification of High-Resolution Images

    Get PDF
    International audienceA Markov random field is a graphical model that is commonly used to combine spectral information and spatial context into image classification problems. The contributions of the spatial versus spectral energies are typically defined by using a smoothing parameter, which is often set empirically. We propose a new framework to estimate the smoothing parameter. For this purpose, we introduce the new concepts of dynamic blocks and class label co-occurrence matrices. The estimation is then based on the analysis of the balance of spatial and spectral energies computed using the spatial class co-occurrence distribution and dynamic blocks. Moreover, we construct a new spatially weighted parameter to preserve the edges, based on the Canny edge detector. We evaluate the performance of the proposed method on three data sets: a multispectral DigitalGlobe WorldView-2 and two hyperspectral images, recorded by the AVIRIS and the ROSIS sensors, respectively. The experimental results show that the proposed method succeeds in estimating the optimal smoothing parameter and yields higher classification accuracies when compared to the state-of-the-art methods

    New Robust Obstacle Detection System Using Color Stereo Vision

    Get PDF
    Intelligent transportation systems (ITS) are divided into intelligent infrastructure systems and intelligent vehicle systems. Intelligent vehicle systems are typically classified in three categories, namely 1) Collision Avoidance Systems; 2) Driver Assistance Systems and 3) Collision Notification Systems. Obstacle detection is one of crucial tasks for Collision Avoidance Systems and Driver Assistance Systems. Obstacle detection systems use vehiclemounted sensors to detect obstuctions, such as other vehicles, bicyclists, pedestrians, road debris, or animals, in a vehicleâs path and alert the driver. Obstacle detection systems are proposed to help drivers see farther and therefore have more time to react to road hazards. These systems also help drivers to get a large visibility area when the visibility conditions is reduced such as night, fog, snow, rain, ... Obstacle detection systems process data acquired from one or several sensors: radar Kruse et al. (2004), lidar Gao & Coifman (2006), monocular vision Lombardi & Zavidovique (2004), stereo vision Franke (2000) Bensrhair et al. (2002) Cabani et al. (2006b) Kogler et al. (2006) Woodfill et al. (2007), vision fused with active sensors Gern et al. (2000) Steux et al. (2002) Mobus & Kolbe (2004)Zhu et al. (2006) Alessandretti et al. (2007)Cheng et al. (2007). It is clear now that most obstacle detection systems cannot work without vision. Typically, vision-based systems consist of cameras that provide gray level images. When visibility conditions are reduced (night, fog, twilight, tunnel, snow, rain), vision systems are almost blind. Obstacle detection systems are less robust and reliable. To deal with the problem of reduced visibility conditions, infrared or color cameras can be used. Thermal imaging cameras are initially used by militaries. Over the last few years, these systems became accessible to the commercial market, and can be found in select 2006 BMW cars. For example, vehicle headlight systems provide between 75 to 140 meters of moderate illumination; at 90 K meters per hour this means less than 4 seconds to react to hazards. When with PathFindIR PathFindIR (n.d.) (a commercial system), a driver can have more than 15 seconds. Other systems still in the research stage assist drivers to detect pedestrians Xu & Fujimura (2002) Broggi et al. (2004) Bertozzi et al. (2007). Color is appropriate to various visibility conditions and various environments. In Betke et al. (2000) and Betke & Nguyen (1998), Betke et al. have demonstrated that the tracking o

    About Edge Detection in Digital Images

    Get PDF
    Edge detection is one of the most commonly used procedures in digital image processing. In the last 30-40 years, many methods and algorithms for edge detection have been proposed. This article presents an overview of edge detection methods, the methods are divided according to the applied basic principles. Next, the measures and image database used for edge detectors performance quantification are described. Ordinary users as well as authors proposing new edge detectors often use Matlab function without understanding it in details. Therefore, one chapter is devoted to some of Matlab function parameters that affect the final result. Finally, the latest trends in edge detection are listed. Picture Lena and two images from Berkeley segmentation data set (BSDS500) are used for edge detection methods comparison

    Gravitation-Based Edge Detection in Hyperspectral Images

    Get PDF
    Edge detection is one of the key issues in the field of computer vision and remote sensing image analysis. Although many different edge-detection methods have been proposed for gray-scale, color, and multispectral images, they still face difficulties when extracting edge features from hyperspectral images (HSIs) that contain a large number of bands with very narrow gap in the spectral domain. Inspired by the clustering characteristic of the gravitational theory, a novel edge-detection algorithm for HSIs is presented in this paper. In the proposed method, we first construct a joint feature space by combining the spatial and spectral features. Each pixel of HSI is assumed to be a celestial object in the joint feature space, which exerts gravitational force to each of its neighboring pixel. Accordingly, each object travels in the joint feature space until it reaches a stable equilibrium. At the equilibrium, the image is smoothed and the edges are enhanced, where the edge pixels can be easily distinguished by calculating the gravitational potential energy. The proposed edge-detection method is tested on several benchmark HSIs and the obtained results were compared with those of four state-of-the-art approaches. The experimental results confirm the efficacy of the proposed method

    Color Image Edge Detection and Segmentation: A Comparison of the Vector Angle and the Euclidean Distance Color Similarity Measures

    Get PDF
    This work is based on Shafer's Dichromatic Reflection Model as applied to color image formation. The color spaces RGB, XYZ, CIELAB, CIELUV, rgb, l1l2l3, and the new h1h2h3 color space are discussed from this perspective. Two color similarity measures are studied: the Euclidean distance and the vector angle. The work in this thesis is motivated from a practical point of view by several shortcomings of current methods. The first problem is the inability of all known methods to properly segment objects from the background without interference from object shadows and highlights. The second shortcoming is the non-examination of the vector angle as a distance measure that is capable of directly evaluating hue similarity without considering intensity especially in RGB. Finally, there is inadequate research on the combination of hue- and intensity-based similarity measures to improve color similarity calculations given the advantages of each color distance measure. These distance measures were used for two image understanding tasks: edge detection, and one strategy for color image segmentation, namely color clustering. Edge detection algorithms using Euclidean distance and vector angle similarity measures as well as their combinations were examined. The list of algorithms is comprised of the modified Roberts operator, the Sobel operator, the Canny operator, the vector gradient operator, and the 3x3 difference vector operator. Pratt's Figure of Merit is used for a quantitative comparison of edge detection results. Color clustering was examined using the k-means (based on the Euclidean distance) and Mixture of Principal Components (based on the vector angle) algorithms. A new quantitative image segmentation evaluation procedure is introduced to assess the performance of both algorithms. Quantitative and qualitative results on many color images (artificial, staged scenes and natural scene images) indicate good edge detection performance using a vector version of the Sobel operator on the h1h2h3 color space. The results using combined hue- and intensity-based difference measures show a slight improvement qualitatively and over using each measure independently in RGB. Quantitative and qualitative results for image segmentation on the same set of images suggest that the best image segmentation results are obtained using the Mixture of Principal Components algorithm on the RGB, XYZ and rgb color spaces. Finally, poor color clustering results in the h1h2h3 color space suggest that some assumptions in deriving a simplified version of the Dichromatic Reflectance Model might have been violated
    corecore