7,890 research outputs found
Broadcasting Convolutional Network for Visual Relational Reasoning
In this paper, we propose the Broadcasting Convolutional Network (BCN) that
extracts key object features from the global field of an entire input image and
recognizes their relationship with local features. BCN is a simple network
module that collects effective spatial features, embeds location information
and broadcasts them to the entire feature maps. We further introduce the
Multi-Relational Network (multiRN) that improves the existing Relation Network
(RN) by utilizing the BCN module. In pixel-based relation reasoning problems,
with the help of BCN, multiRN extends the concept of `pairwise relations' in
conventional RNs to `multiwise relations' by relating each object with multiple
objects at once. This yields in O(n) complexity for n objects, which is a vast
computational gain from RNs that take O(n^2). Through experiments, multiRN has
achieved a state-of-the-art performance on CLEVR dataset, which proves the
usability of BCN on relation reasoning problems.Comment: Accepted paper at ECCV 2018. 24 page
Convolutional Sparse Kernel Network for Unsupervised Medical Image Analysis
The availability of large-scale annotated image datasets and recent advances
in supervised deep learning methods enable the end-to-end derivation of
representative image features that can impact a variety of image analysis
problems. Such supervised approaches, however, are difficult to implement in
the medical domain where large volumes of labelled data are difficult to obtain
due to the complexity of manual annotation and inter- and intra-observer
variability in label assignment. We propose a new convolutional sparse kernel
network (CSKN), which is a hierarchical unsupervised feature learning framework
that addresses the challenge of learning representative visual features in
medical image analysis domains where there is a lack of annotated training
data. Our framework has three contributions: (i) We extend kernel learning to
identify and represent invariant features across image sub-patches in an
unsupervised manner. (ii) We initialise our kernel learning with a layer-wise
pre-training scheme that leverages the sparsity inherent in medical images to
extract initial discriminative features. (iii) We adapt a multi-scale spatial
pyramid pooling (SPP) framework to capture subtle geometric differences between
learned visual features. We evaluated our framework in medical image retrieval
and classification on three public datasets. Our results show that our CSKN had
better accuracy when compared to other conventional unsupervised methods and
comparable accuracy to methods that used state-of-the-art supervised
convolutional neural networks (CNNs). Our findings indicate that our
unsupervised CSKN provides an opportunity to leverage unannotated big data in
medical imaging repositories.Comment: Accepted by Medical Image Analysis (with a new title 'Convolutional
Sparse Kernel Network for Unsupervised Medical Image Analysis'). The
manuscript is available from following link
(https://doi.org/10.1016/j.media.2019.06.005
The Effectiveness of Data Augmentation for Detection of Gastrointestinal Diseases from Endoscopical Images
The lack, due to privacy concerns, of large public databases of medical
pathologies is a well-known and major problem, substantially hindering the
application of deep learning techniques in this field. In this article, we
investigate the possibility to supply to the deficiency in the number of data
by means of data augmentation techniques, working on the recent Kvasir dataset
of endoscopical images of gastrointestinal diseases. The dataset comprises
4,000 colored images labeled and verified by medical endoscopists, covering a
few common pathologies at different anatomical landmarks: Z-line, pylorus and
cecum. We show how the application of data augmentation techniques allows to
achieve sensible improvements of the classification with respect to previous
approaches, both in terms of precision and recall
Automatic Labeled LiDAR Data Generation based on Precise Human Model
Following improvements in deep neural networks, state-of-the-art networks
have been proposed for human recognition using point clouds captured by LiDAR.
However, the performance of these networks strongly depends on the training
data. An issue with collecting training data is labeling. Labeling by humans is
necessary to obtain the ground truth label; however, labeling requires huge
costs. Therefore, we propose an automatic labeled data generation pipeline, for
which we can change any parameters or data generation environments. Our
approach uses a human model named Dhaiba and a background of Miraikan and
consequently generated realistic artificial data. We present 500k+ data
generated by the proposed pipeline. This paper also describes the specification
of the pipeline and data details with evaluations of various approaches.Comment: Accepted at ICRA201
Clinical Assistant Diagnosis for Electronic Medical Record Based on Convolutional Neural Network
Automatically extracting useful information from electronic medical records
along with conducting disease diagnoses is a promising task for both clinical
decision support(CDS) and neural language processing(NLP). Most of the existing
systems are based on artificially constructed knowledge bases, and then
auxiliary diagnosis is done by rule matching. In this study, we present a
clinical intelligent decision approach based on Convolutional Neural
Networks(CNN), which can automatically extract high-level semantic information
of electronic medical records and then perform automatic diagnosis without
artificial construction of rules or knowledge bases. We use collected 18,590
copies of the real-world clinical electronic medical records to train and test
the proposed model. Experimental results show that the proposed model can
achieve 98.67\% accuracy and 96.02\% recall, which strongly supports that using
convolutional neural network to automatically learn high-level semantic
features of electronic medical records and then conduct assist diagnosis is
feasible and effective.Comment: 9 pages, 4 figures, Accepted by Scientific Report
- …