1,656 research outputs found
Large-Scale Study of Perceptual Video Quality
The great variations of videographic skills, camera designs, compression and
processing protocols, and displays lead to an enormous variety of video
impairments. Current no-reference (NR) video quality models are unable to
handle this diversity of distortions. This is true in part because available
video quality assessment databases contain very limited content, fixed
resolutions, were captured using a small number of camera devices by a few
videographers and have been subjected to a modest number of distortions. As
such, these databases fail to adequately represent real world videos, which
contain very different kinds of content obtained under highly diverse imaging
conditions and are subject to authentic, often commingled distortions that are
impossible to simulate. As a result, NR video quality predictors tested on
real-world video data often perform poorly. Towards advancing NR video quality
prediction, we constructed a large-scale video quality assessment database
containing 585 videos of unique content, captured by a large number of users,
with wide ranges of levels of complex, authentic distortions. We collected a
large number of subjective video quality scores via crowdsourcing. A total of
4776 unique participants took part in the study, yielding more than 205000
opinion scores, resulting in an average of 240 recorded human opinions per
video. We demonstrate the value of the new resource, which we call the LIVE
Video Quality Challenge Database (LIVE-VQC), by conducting a comparison of
leading NR video quality predictors on it. This study is the largest video
quality assessment study ever conducted along several key dimensions: number of
unique contents, capture devices, distortion types and combinations of
distortions, study participants, and recorded subjective scores. The database
is available for download on this link:
http://live.ece.utexas.edu/research/LIVEVQC/index.html
Recommended from our members
Subjective and objective quality evaluation of synthetic and high dynamic range images
Recent years have seen a huge growth in the acquisition, transmission, and storage of videos. The visual data consists of both natural scenes as well as synthetic scenes, such as animated movies, cartoons and video games. In all these cases, the ultimate goal is to provide the viewers with a satisfactory quality-of-experience. In addition to the traditional 8-bit images, high dynamic range imaging is also becoming popular because of its ability to represent the real world luminances more realistically. Coming up with objective image quality assessment algorithms for these applications is an interesting research problem. In this work, I have developed a synthetic image quality database by introducing varying degrees of different types of distortions and conducted a subjective experiment in order to obtain the ground-truth data. I evaluated the performance of state-of-the-art image quality assessment algorithms (typically meant for natural images) on this database, especially no-reference algorithms that have not been applied to the domain of computer graphics images before. I identified the top-performing algorithms along with analyzing the types of distortions on which the present algorithms show a less impressive performance. For high dynamic range(HDR) images, I have designed two new full-reference image quality assessment algorithms to judge the quality of tonemapped HDR images using statistical features extracted from them. I have also conducted a massive online crowd-sourced subjective test for HDR image artifacts arising from tonemapping, multiple-exposure fusion and post processing. To the best of our knowledge, presently this is the largest HDR image database in the world involving the largest number of source images and most number of human evaluations. Based on the subjective evaluations obtained, I have also proposed machine learning based no-reference image quality assessment algorithms to predict the perceptual quality of HDR images.Electrical and Computer Engineerin
Virtual Globes for UAV-based data integration: Sputnik GIS and Google Earth™ applications
“This is an Accepted Manuscript of an article published by Taylor & Francis in International Journal of Digital Earth on 03 May 2018, available online: https://www.tandfonline.com/doi/abs/10.1080/17538947.2018.1470205"The integration of local measurements and monitoring via global-scale Earth
observations has become a new challenge in digital Earth science. The increasing
accessibility and ease of use of virtual globes (VGs) represent primary advantages of
this integration, and the digital Earth scientific community has adopted this
technology as one of the main methods for disseminating the results of scientific
studies. In this study, the best VG software for the dissemination and analysis of
high-resolution UAV (Unmanned Aerial Vehicle) data is identified for global and
continuous geographic scope support. The VGs Google Earth and Sputnik
Geographic Information System (GIS) are selected and compared for this purpose.
Google Earth is a free platform and one of the most widely used VGs, and one of its
best features its ability to provide users with quality visual results. The proprietary
software Sputnik GIS more closely approximates the analytical capacity of a
traditional GIS and provides outstanding advantages, such as DEM overlapping and
visualization for its disseminationThis work was supported by Xunta de Galicia under the Grant “Financial aid for the consolidation and structure of competitive units of investigation in the universities of the University Galician System (2016-18)” (Ref. ED431B 2016/030 and Ref. ED341D R2016/023). The authors also acknowledge support provided by “Realización de vuelos virtuales en las parcelas del proyecto Green deserts LIFE09 / ENV/ES / 000447”S
Recommended from our members
Perceptual quality assessment of real-world images and videos
The development of online social-media venues and rapid advances in technology by camera and mobile device manufacturers have led to the creation and consumption of a seemingly limitless supply of visual content. However, a vast majority of these digital images and videos are often afflicted with annoying artifacts during acquisition, subsequent storage, and transmission over the network. All these factors impact the quality of the visual media as perceived by a human observer, thereby compromising their quality of experience (QoE).
This dissertation focuses on constructing datasets that are representative of real-world image and video distortions as well as on designing algorithms that accurately predict the perceptual quality of images and videos. The primary goal of this research is to design and demonstrate automatic image and continuous-time video quality predictors that can effectively tackle the widely diverse authentic spatial, temporal, and network-induced distortions -- contrary to all present-day algorithms that operate on single, synthetic visual distortions and predict a single overall quality score for a given video.
I introduce an image quality database which contains a large number of images captured using a representative variety of modern mobile devices and afflicted with a widely diverse authentic image distortions. I will also describe the design of an online crowdsourcing system which aided a very large-scale image quality assessment subjective study. This data collection facilitated the design of a new image quality predictor that is founded on the principles of natural scene statistics of images in different color spaces and transform domains. This new quality method is capable of assessing the quality of images with complex mixtures of distortions and yields high correlation with human perception.
Pertaining to videos, this dissertation describes a video quality database created to understand the impact of network-induced distortions on an end user's quality of experience. I present the details of a large-scale subjective study that I conducted to gather continuous-time ground truth QoE scores on a collection of 180 videos afflicted with diverse stalling events. I also present my analysis of the temporal variations in the perceived QoE due to the time-varying video quality and present insights on the impact of relevant human cognitive aspects such as long-term and short-term memory and recency on quality perception. Next, I present a continuous-time objective QoE predicting model that effectively captures the complex interactions between the aforementioned human cognitive elements, spatial and temporal distortions, properties of stalling events, and models the state of any given client-side network buffer. I also show how the proposed framework can be extended by further supplementing with any number of additional inputs (or by eliminating any ineffective ones), based on the information available at the content providers during the design of adaptive stream-switching algorithms. This QoE predictor supports future research in the design of quality-aware stream-switching algorithms which could control the position, location, and length of stalls, given a network bandwidth budget and the end user's device information, such that the end user's QoE is maximized.Computer Science
Crowdsourced intuitive visual design feedback
For many people images are a medium preferable to text and yet, with the exception of
star ratings, most formats for conventional computer mediated feedback focus on text.
This thesis develops a new method of crowd feedback for designers based on images.
Visual summaries are generated from a crowd’s feedback images chosen in response to
a design. The summaries provide the designer with impressionistic and inspiring visual
feedback. The thesis sets out the motivation for this new method, describes the
development of perceptually organised image sets and a summarisation algorithm to
implement it. Evaluation studies are reported which, through a mixed methods
approach, provide evidence of the validity and potential of the new image-based
feedback method.
It is concluded that the visual feedback method would be more appealing than text for
that section of the population who may be of a visual cognitive style. Indeed the
evaluation studies are evidence that such users believe images are as good as text when
communicating their emotional reaction about a design. Designer participants reported
being inspired by the visual feedback where, comparably, they were not inspired by
text. They also reported that the feedback can represent the perceived mood in their
designs, and that they would be enthusiastic users of a service offering this new form of
visual design feedback
- …