10 research outputs found
Multimodal Multipart Learning for Action Recognition in Depth Videos
The articulated and complex nature of human actions makes the task of action
recognition difficult. One approach to handle this complexity is dividing it to
the kinetics of body parts and analyzing the actions based on these partial
descriptors. We propose a joint sparse regression based learning method which
utilizes the structured sparsity to model each action as a combination of
multimodal features from a sparse set of body parts. To represent dynamics and
appearance of parts, we employ a heterogeneous set of depth and skeleton based
features. The proper structure of multimodal multipart features are formulated
into the learning framework via the proposed hierarchical mixed norm, to
regularize the structured features of each part and to apply sparsity between
them, in favor of a group feature selection. Our experimental results expose
the effectiveness of the proposed learning method in which it outperforms other
methods in all three tested datasets while saturating one of them by achieving
perfect accuracy
Hierarchical visual perception and two-dimensional compressive sensing for effective content-based color image retrieval
Content-based image retrieval (CBIR) has been an active research theme in the computer vision community for over two decades. While the field is relatively mature, significant research is still required in this area to develop solutions for practical applications. One reason that practical solutions have not yet been realized could be due to a limited understanding of the cognitive aspects of the human vision system. Inspired by three cognitive properties of human vision, namely, hierarchical structuring, color perception and embedded compressive sensing, a new CBIR approach is proposed. In the proposed approach, the Hue, Saturation and Value (HSV) color model and the Similar Gray Level Co-occurrence Matrix (SGLCM) texture descriptors are used to generate elementary features. These features then form a hierarchical representation of the data to which a two-dimensional compressive sensing (2D CS) feature mining algorithm is applied. Finally, a weighted feature matching method is used to perform image retrieval. We present a comprehensive set of results of applying our proposed Hierarchical Visual Perception Enabled 2D CS approach using publicly available datasets and demonstrate the efficacy of our techniques when compared with other recently published, state-of-the-art approaches
Broad Learning for Healthcare
A broad spectrum of data from different modalities are generated in the
healthcare domain every day, including scalar data (e.g., clinical measures
collected at hospitals), tensor data (e.g., neuroimages analyzed by research
institutes), graph data (e.g., brain connectivity networks), and sequence data
(e.g., digital footprints recorded on smart sensors). Capability for modeling
information from these heterogeneous data sources is potentially transformative
for investigating disease mechanisms and for informing therapeutic
interventions.
Our works in this thesis attempt to facilitate healthcare applications in the
setting of broad learning which focuses on fusing heterogeneous data sources
for a variety of synergistic knowledge discovery and machine learning tasks. We
are generally interested in computer-aided diagnosis, precision medicine, and
mobile health by creating accurate user profiles which include important
biomarkers, brain connectivity patterns, and latent representations. In
particular, our works involve four different data mining problems with
application to the healthcare domain: multi-view feature selection, subgraph
pattern mining, brain network embedding, and multi-view sequence prediction.Comment: PhD Thesis, University of Illinois at Chicago, March 201