3,569 research outputs found
Improving utility of brain tumor confocal laser endomicroscopy: objective value assessment and diagnostic frame detection with convolutional neural networks
Confocal laser endomicroscopy (CLE), although capable of obtaining images at
cellular resolution during surgery of brain tumors in real time, creates as
many non-diagnostic as diagnostic images. Non-useful images are often distorted
due to relative motion between probe and brain or blood artifacts. Many images,
however, simply lack diagnostic features immediately informative to the
physician. Examining all the hundreds or thousands of images from a single case
to discriminate diagnostic images from nondiagnostic ones can be tedious.
Providing a real-time diagnostic value assessment of images (fast enough to be
used during the surgical acquisition process and accurate enough for the
pathologist to rely on) to automatically detect diagnostic frames would
streamline the analysis of images and filter useful images for the
pathologist/surgeon. We sought to automatically classify images as diagnostic
or non-diagnostic. AlexNet, a deep-learning architecture, was used in a 4-fold
cross validation manner. Our dataset includes 16,795 images (8572 nondiagnostic
and 8223 diagnostic) from 74 CLE-aided brain tumor surgery patients. The ground
truth for all the images is provided by the pathologist. Average model accuracy
on test data was 91% overall (90.79 % accuracy, 90.94 % sensitivity and 90.87 %
specificity). To evaluate the model reliability we also performed receiver
operating characteristic (ROC) analysis yielding 0.958 average for the area
under ROC curve (AUC). These results demonstrate that a deeply trained AlexNet
network can achieve a model that reliably and quickly recognizes diagnostic CLE
images.Comment: SPIE Medical Imaging: Computer-Aided Diagnosis 201
Deep Neural Networks for No-Reference and Full-Reference Image Quality Assessment
We present a deep neural network-based approach to image quality assessment
(IQA). The network is trained end-to-end and comprises ten convolutional layers
and five pooling layers for feature extraction, and two fully connected layers
for regression, which makes it significantly deeper than related IQA models.
Unique features of the proposed architecture are that: 1) with slight
adaptations it can be used in a no-reference (NR) as well as in a
full-reference (FR) IQA setting and 2) it allows for joint learning of local
quality and local weights, i.e., relative importance of local quality to the
global quality estimate, in an unified framework. Our approach is purely
data-driven and does not rely on hand-crafted features or other types of prior
domain knowledge about the human visual system or image statistics. We evaluate
the proposed approach on the LIVE, CISQ, and TID2013 databases as well as the
LIVE In the wild image quality challenge database and show superior performance
to state-of-the-art NR and FR IQA methods. Finally, cross-database evaluation
shows a high ability to generalize between different databases, indicating a
high robustness of the learned features
Deep CNN Model for Non-Screen Content and Screen Content Image Quality Assessment
In the current world, user experience in various platforms matters a lot for different organizations. But providing a better experience can be challenging if the multimedia content on online platforms is having different kinds of distortions which impact the overall experience of the user. There can be various reasons behind distortions such as compression or minimal lighting condition while taking photos. In this work, a deep CNN-based Non-Screen Content and Screen Content NR-IQA framework is proposed which solves this issue in a more effective way. The framework is known as DNSSCIQ. Two different architectures are proposed based upon the input image type whether the input is a screen content or non-screen content image. This work attempts to solve this by evaluating the quality of such image
- …