10 research outputs found
CapProNet: Deep Feature Learning via Orthogonal Projections onto Capsule Subspaces
In this paper, we formalize the idea behind capsule nets of using a capsule
vector rather than a neuron activation to predict the label of samples. To this
end, we propose to learn a group of capsule subspaces onto which an input
feature vector is projected. Then the lengths of resultant capsules are used to
score the probability of belonging to different classes. We train such a
Capsule Projection Network (CapProNet) by learning an orthogonal projection
matrix for each capsule subspace, and show that each capsule subspace is
updated until it contains input feature vectors corresponding to the associated
class. We will also show that the capsule projection can be viewed as
normalizing the multiple columns of the weight matrix simultaneously to form an
orthogonal basis, which makes it more effective in incorporating novel
components of input features to update capsule representations. In other words,
the capsule projection can be viewed as a multi-dimensional weight
normalization in capsule subspaces, where the conventional weight normalization
is simply a special case of the capsule projection onto 1D lines. Only a small
negligible computing overhead is incurred to train the network in
low-dimensional capsule subspaces or through an alternative hyper-power
iteration to estimate the normalization matrix. Experiment results on image
datasets show the presented model can greatly improve the performance of the
state-of-the-art ResNet backbones by and that of the Densenet by
respectively at the same level of computing and memory expenses. The
CapProNet establishes the competitive state-of-the-art performance for the
family of capsule nets by significantly reducing test errors on the benchmark
datasets.Comment: Liheng Zhang, Marzieh Edraki, Guo-Jun Qi. CapProNet: Deep Feature
Learning via Orthogonal Projections onto Capsule Subspaces, in Proccedings of
Thirty-second Conference on Neural Information Processing Systems (NIPS
2018), Palais des Congr\`es de Montr\'eal, Montr\'eal, Canda, December 3-8,
201
Implication of Manifold Assumption in Deep Learning Models for Computer Vision Applications
The Deep Neural Networks (DNN) have become the main contributor in the field of machine learning (ML). Specifically in the computer vision (CV), there are applications like image and video classification, object detection and tracking, instance segmentation and visual question answering, image and video generation are some of the applications from many that DNNs have demonstrated magnificent progress. To achieve the best performance, the DNNs usually require a large number of labeled samples, and finding the optimal solution for such complex models with millions of parameters is a challenging task. It is known that, the data are not uniformly distributed on the sample space, rather they are residing on a low-dimensional manifold embedded in the ambient space. In this dissertation, we specifically investigate the effect of manifold assumption on various applications in computer vision. First we propose a novel loss sensitive adversarial learning (LSAL) paradigm in training GAN framework that is built upon the assumption that natural images are lying on a smooth manifold. It benefits from the geodesic of samples in addition to the distance of samples in the ambient space to differentiate between real and generated samples. It is also shown that the discriminator of a GAN model trained based on LSAL paradigm is also successful in semi-supervised classification of images when the number of labeled images are limited. Then we propose a novel Capsule projection Network (CapProNet) that models the manifold of data through the union of subspace capsules in the last layer of a CNN image classifier. The CapProNet idea has been further extended to the general framework of Subspace Capsule Network that not only does model the deformation of objects but also parts of objects through the hierarchy of sub- space capsules layers. We apply the subspace capsule network on the tasks of (semi-) supervised image classification and also high resolution image generation. Finally, we verify the reliability of DNN models by investigating the intrinsic properties of the models around the manifold of data to detect maliciously trained Trojan models
Global versus Localized Generative Adversarial Nets
In this paper, we present a novel localized Generative Adversarial Net (GAN)
to learn on the manifold of real data. Compared with the classic GAN that {\em
globally} parameterizes a manifold, the Localized GAN (LGAN) uses local
coordinate charts to parameterize distinct local geometry of how data points
can transform at different locations on the manifold. Specifically, around each
point there exists a {\em local} generator that can produce data following
diverse patterns of transformations on the manifold. The locality nature of
LGAN enables local generators to adapt to and directly access the local
geometry without need to invert the generator in a global GAN. Furthermore, it
can prevent the manifold from being locally collapsed to a dimensionally
deficient tangent subspace by imposing an orthonormality prior between
tangents. This provides a geometric approach to alleviating mode collapse at
least locally on the manifold by imposing independence between data
transformations in different tangent directions. We will also demonstrate the
LGAN can be applied to train a robust classifier that prefers locally
consistent classification decisions on the manifold, and the resultant
regularizer is closely related with the Laplace-Beltrami operator. Our
experiments show that the proposed LGANs can not only produce diverse image
transformations, but also deliver superior classification performances
SubSpace Capsule Network
Convolutional neural networks (CNNs) have become a key asset to most of fields in AI. Despite their successful performance, CNNs suffer from a major drawback. They fail to capture the hierarchy of spatial relation among different parts of an entity. As a remedy to this problem, the idea of capsules was proposed by Hinton. In this paper, we propose the SubSpace Capsule Network (SCN) that exploits the idea of capsule networks to model possible variations in the appearance or implicitly-defined properties of an entity through a group of capsule subspaces instead of simply grouping neurons to create capsules. A capsule is created by projecting an input feature vector from a lower layer onto the capsule subspace using a learnable transformation. This transformation finds the degree of alignment of the input with the properties modeled by the capsule subspace.We show that SCN is a general capsule network that can successfully be applied to both discriminative and generative models without incurring computational overhead compared to CNN during test time. Effectiveness of SCN is evaluated through a comprehensive set of experiments on supervised image classification, semi-supervised image classification and high-resolution image generation tasks using the generative adversarial network (GAN) framework. SCN significantly improves the performance of the baseline models in all 3 tasks
Cappronet: Deep Feature Learning Via Orthogonal Projections Onto Capsule Subspaces
In this paper, we formalize the idea behind capsule nets of using a capsule vector rather than a neuron activation to predict the label of samples. To this end, we propose to learn a group of capsule subspaces onto which an input feature vector is projected. Then the lengths of resultant capsules are used to score the probability of belonging to different classes. We train such a Capsule Projection Network (CapProNet) by learning an orthogonal projection matrix for each capsule subspace, and show that each capsule subspace is updated until it contains input feature vectors corresponding to the associated class. We will also show that the capsule projection can be viewed as normalizing the multiple columns of the weight matrix simultaneously to form an orthogonal basis, which makes it more effective in incorporating novel components of input features to update capsule representations. In other words, the capsule projection can be viewed as a multi-dimensional weight normalization in capsule subspaces, where the conventional weight normalization is simply a special case of the capsule projection onto 1D lines. Only a small negligible computing overhead is incurred to train the network in low-dimensional capsule subspaces or through an alternative hyper-power iteration to estimate the normalization matrix. Experiment results on image datasets show the presented model can greatly improve the performance of the state-of-the-art ResNet backbones by 10 − 20% and that of the Densenet by 5 − 7% respectively at the same level of computing and memory expenses. The CapProNet establishes the competitive state-of-the-art performance for the family of capsule nets by significantly reducing test errors on the benchmark datasets
Relationship between Mothers’ Spiritual Health Scores with Newborns’ Physical Development Indices and Physiologic Parameters in Hazrat Zeinab Training Hospital
Background: Spiritual health is one of the important factors predicting human health. This study aimed to determine the relationship between mothers’ spiritual health with newborns’ physical development indices and other physiologic parameters.Methods: In this cross-sectional study, 155 mothers giving birth to newborns were selected from Hazrat Zeinab hospital during 2017-2018. The data were gathered through a checklist containing all mothers’ and babies’ demographic information. Moreover, we used Palutzian and Ellison’s scale to measure the mothers’ spiritual health score. To analyze the data, we used SPSS software (version 18).Results: The mothers’ mean age was reported as 27.84±6.67 years. Moreover, 71.6% of the mothers’ educational level was under diploma, and 65.2% of them were not employed (did not have any jobs). The mean score of the mothers’ spiritual health was 75.96±8.75. In this regard, 97.4% of the subjects had a moderate level of spiritual health, and 2.6% of them had a high level of spiritual health. There was a significant negative correlation between the mothers’ spiritual health score and neonates’ physical development scores. However, this correlation was significant (height: r=-0.1, P=0.21; weight r=-0.058, P=0.47; size of head: r=-0.033, P=0.6; size of belly: r=0.047, P=0.56), and there were positive correlations between the mothers’ spiritual health scores (heart beats: r=-0.034, P=0.66; percentage of saturated oxygen: r=-0.034, P=0.90; degree of heat: r=0.047, P=0.96). However, none of these correlations were statistically significant.Conclusion: In general, the results of this study showed that most of the mothers had a normal and high level of spiritual health, but a higher percentage of moderate level of spiritual health was observed in mothers, compared to those of other levels. Moreover, no significant correlations were found between mothers’ spiritual health scores with newborns’ physical development indices and other physiologic factors
Global Versus Localized Generative Adversarial Nets
In this paper, we present a novel localized Generative Adversarial Net (GAN) to learn on the manifold of real data. Compared with the classic GAN that globally parameterizes a manifold, the Localized GAN (LGAN) uses local coordinate charts to parameterize distinct local geometry of how data points can transform at different locations on the manifold. Specifically, around each point there exists a local generator that can produce data following diverse patterns of transformations on the manifold. The locality nature of LGAN enables local generators to adapt to and directly access the local geometry without need to invert the generator in a global GAN. Furthermore, it can prevent the manifold from being locally collapsed to a dimensionally deficient tangent subspace by imposing an orthonormality prior between tangents. This provides a geometric approach to alleviating mode collapse at least locally on the manifold by imposing independence between data transformations in different tangent directions. We will also demonstrate the LGAN can be applied to train a robust classifier that prefers locally consistent classification decisions on the manifold, and the resultant regularizer is closely related with the Laplace-Beltrami operator. Our experiments show that the proposed LGANs can not only produce diverse image transformations, but also deliver superior classification performances
The Effect of Instructing the Principles of Endotracheal Tube Suctioning on Knowledge and Performance of Nursing Staff Working in Neonatal Intensive Care Units in Shiraz University of Medical Sciences
Introduction: Nurses must be aware of the risks regarding endotracheal tube suctioning and should have continuing education in this field. This study was performed to assess the impact of instruction on the knowledge and performance of NICU nursing staff in Shiraz University of Medical Sciences in 2006.
Methods: Fifty nurses of neonatal intensive care units participated in this quasi experimental study. At first, their knowledge and performance in neonatal endotracheal tube suctioning was investigated using test and checklist. After specifying the experimental and control group through systematic random allocation, the suctioning instruction was done for experimental group and infection prevention instruction was done for control group. Two days and 2 months after instruction, nurses' knowledge and performance were assessed again. Data analysis was done using Chi- Square, Mann Whitney, and Wilcoxon by SPSS software.
Results: The means for knowledge and performance of experimental group respectively two days and two months after instruction was 16.56 and arrived from this score to 28.48 and 27.4 and from 20.6 arrived to 39.14 and 38.34.
Conclusion: Instructing the principles of endotracheal tube suctioning improves the level of knowledge and performance in nurses. Since education effect declines gradually, continuing education in this field seems to be necessary
An Adversarial Approach To Hard Triplet Generation
While deep neural networks have demonstrated competitive results for many visual recognition and image retrieval tasks, the major challenge lies in distinguishing similar images from different categories (i.e., hard negative examples) while clustering images with large variations from the same category (i.e., hard positive examples). The current state-of-the-art is to mine the most hard triplet examples from the mini-batch to train the network. However, mining-based methods tend to look into these triplets that are hard in terms of the current estimated network, rather than deliberately generating those hard triplets that really matter in globally optimizing the network. For this purpose, we propose an adversarial network for Hard Triplet Generation (HTG) to optimize the network ability in distinguishing similar examples of different categories as well as grouping varied examples of the same categories. We evaluate our method on the real-world challenging datasets, such as CUB200-2011, CARS196, DeepFashion and VehicleID datasets, and show that our method outperforms the state-of-the-art methods significantly