4,466 research outputs found
Representations and representation learning for image aesthetics prediction and image enhancement
With the continual improvement in cell phone cameras and improvements in the connectivity of mobile devices, we have seen an exponential increase in the images that are captured, stored and shared on social media. For example, as of July 1st 2017 Instagram had over 715 million registered users which had posted just shy of 35 billion images. This represented approximately seven and nine-fold increase in the number of users and photos present on Instagram since 2012. Whether the images are stored on personal computers or reside on social networks (e.g. Instagram, Flickr), the sheer number of images calls for methods to determine various image properties, such as object presence or appeal, for the purpose of automatic image management and curation. One of the central problems in consumer photography centers around determining the aesthetic appeal of an image and motivates us to explore questions related to understanding aesthetic preferences, image enhancement and the possibility of using such models on devices with constrained resources.
In this dissertation, we present our work on exploring representations and representation learning approaches for aesthetic inference, composition ranking and its application to image enhancement. Firstly, we discuss early representations that mainly consisted of expert features, and their possibility to enhance Convolutional Neural Networks (CNN). Secondly, we discuss the ability of resource-constrained CNNs, and the different architecture choices (inputs size and layer depth) in solving various aesthetic inference tasks: binary classification, regression, and image cropping. We show that if trained for solving fine-grained aesthetics inference, such models can rival the cropping performance of other aesthetics-based croppers, however they fall short in comparison to models trained for composition ranking. Lastly, we discuss our work on exploring and identifying the design choices in training composition ranking functions, with the goal of using them for image composition enhancement
Image Aesthetic Assessment: A Comparative Study of Hand-Crafted & Deep Learning Models
publishedVersio
Towards Artistic Image Aesthetics Assessment: a Large-scale Dataset and a New Method
Image aesthetics assessment (IAA) is a challenging task due to its highly
subjective nature. Most of the current studies rely on large-scale datasets
(e.g., AVA and AADB) to learn a general model for all kinds of photography
images. However, little light has been shed on measuring the aesthetic quality
of artistic images, and the existing datasets only contain relatively few
artworks. Such a defect is a great obstacle to the aesthetic assessment of
artistic images. To fill the gap in the field of artistic image aesthetics
assessment (AIAA), we first introduce a large-scale AIAA dataset: Boldbrush
Artistic Image Dataset (BAID), which consists of 60,337 artistic images
covering various art forms, with more than 360,000 votes from online users. We
then propose a new method, SAAN (Style-specific Art Assessment Network), which
can effectively extract and utilize style-specific and generic aesthetic
information to evaluate artistic images. Experiments demonstrate that our
proposed approach outperforms existing IAA methods on the proposed BAID dataset
according to quantitative comparisons. We believe the proposed dataset and
method can serve as a foundation for future AIAA works and inspire more
research in this field. Dataset and code are available at:
https://github.com/Dreemurr-T/BAID.gitComment: Accepted by CVPR 202
- …