15 research outputs found
MEGAN: Mixture of Experts of Generative Adversarial Networks for Multimodal Image Generation
Recently, generative adversarial networks (GANs) have shown promising
performance in generating realistic images. However, they often struggle in
learning complex underlying modalities in a given dataset, resulting in
poor-quality generated images. To mitigate this problem, we present a novel
approach called mixture of experts GAN (MEGAN), an ensemble approach of
multiple generator networks. Each generator network in MEGAN specializes in
generating images with a particular subset of modalities, e.g., an image class.
Instead of incorporating a separate step of handcrafted clustering of multiple
modalities, our proposed model is trained through an end-to-end learning of
multiple generators via gating networks, which is responsible for choosing the
appropriate generator network for a given condition. We adopt the categorical
reparameterization trick for a categorical decision to be made in selecting a
generator while maintaining the flow of the gradients. We demonstrate that
individual generators learn different and salient subparts of the data and
achieve a multiscale structural similarity (MS-SSIM) score of 0.2470 for CelebA
and a competitive unsupervised inception score of 8.33 in CIFAR-10.Comment: 27th International Joint Conference on Artificial Intelligence (IJCAI
2018
F^2-Softmax: Diversifying Neural Text Generation via Frequency Factorized Softmax
Despite recent advances in neural text generation, encoding the rich
diversity in human language remains elusive. We argue that the sub-optimal text
generation is mainly attributable to the imbalanced token distribution, which
particularly misdirects the learning model when trained with the
maximum-likelihood objective. As a simple yet effective remedy, we propose two
novel methods, F^2-Softmax and MefMax, for a balanced training even with the
skewed frequency distribution. MefMax assigns tokens uniquely to frequency
classes, trying to group tokens with similar frequencies and equalize frequency
mass between the classes. F^2-Softmax then decomposes a probability
distribution of the target token into a product of two conditional
probabilities of (i) frequency class, and (ii) token from the target frequency
class. Models learn more uniform probability distributions because they are
confined to subsets of vocabularies. Significant performance gains on seven
relevant metrics suggest the supremacy of our approach in improving not only
the diversity but also the quality of generated texts.Comment: EMNLP 202
Recommended from our members
Precise Identification of Neurological Disorders using Deep Learning and Multimodal Clinical Neuroimaging
Neurological disorders present a significant challenge in global health. With the increasing availability of imaging datasets and the development of precise machine learning models, early and accurate diagnosis of neurological conditions is a promising and active area of research. However, several characteristic factors in neurology domains, such as heterogeneous imaging, inaccurate labels, or limited data, act as bottlenecks in using deep learning on clinical neuroimaging.
Given these circumstances, this dissertation attempts to provide a guideline, proposing several methods and showcasing successful implementations in broad neurological conditions, including epilepsy and neurodegeneration. Methodologically, a particular focus is on comparing a two-dimensional approach as opposed to three-dimensional neural networks. In most clinical domains of neurological disorders, data are scarce and signals are weak, discouraging the use of 3D representation of raw scan data. This dissertation first demonstrates competitive performances with 2D models in tuber segmentation and AD comorbidity detection.
Second, the potentials of ensemble learning are explored, further justifying the use of 2D models in the identification of neurodegeneration. Lastly, CleanNeuro is introduced in the context of 2D classification, a novel algorithm for denoising the datasets prior to training. CleanNeuro, on top of 2D classification and ensemble learning, demonstrates the feasibility of accurately classifying patients with comorbid AD and cerebral amyloid angiopathy from AD controls. Methods presented in this dissertation may serve as exemplars in the study of neurological disorders using deep learning and clinical neuroimaging.
Clinically, this dissertation contributes to improving automated diagnosis and identification of regional vulnerabilities of several neurological disorders on clinical neuroimaging using deep learning. First, the classification of patients with Alzheimer’s disease from cognitively normal group demonstrates the potentials of using positron emission tomography with tau tracers as a competitive biomarker for precision medicine. Second, the segmentation of tubers in patients with tuberous sclerosis complex proves a successful 2D modeling approach in quantifying neurological burden of a rare yet deadly disease. Third, the detection of comorbid pathologies from patients with Alzheimer’s disease is analyzed and discussed in depth. Based on prior findings that comorbidities of Alzheimer’s disease affect the brain structure in a distinctive pattern, this dissertation proves for the first time the effectiveness of using deep learning on the accurate identification of comorbid pathology in vivo. Leveraging postmortem neuropathology as ground truth labels on top of the proposed methods records competitive performances in comorbidity prediction. Notably, this dissertation discovers that structural magnetic resonance imaging is a reliable biomarker in differentiating the comorbid cereberal amyloid angiopathy from Alzheimer’s disease patients.
The dissertation discusses experimental findings on a wide range of neurological disorders, including tuberous sclerosis complex, dementia, and epilepsy. These results contribute to better decision-making on building neural network models for understanding and managing neurological diseases. With the thorough exploration, the dissertation may provide valuable insights that can push forward research in clinical neurology
Overestimated prediction using polygenic prediction derived from summary statistics
Abstract Background When polygenic risk score (PRS) is derived from summary statistics, independence between discovery and test sets cannot be monitored. We compared two types of PRS studies derived from raw genetic data (denoted as rPRS) and the summary statistics for IGAP (sPRS). Results Two variables with the high heritability in UK Biobank, hypertension, and height, are used to derive an exemplary scale effect of PRS. sPRS without APOE is derived from International Genomics of Alzheimer’s Project (IGAP), which records ΔAUC and ΔR2 of 0.051 ± 0.013 and 0.063 ± 0.015 for Alzheimer’s Disease Sequencing Project (ADSP) and 0.060 and 0.086 for Accelerating Medicine Partnership - Alzheimer’s Disease (AMP-AD). On UK Biobank, rPRS performances for hypertension assuming a similar size of discovery and test sets are 0.0036 ± 0.0027 (ΔAUC) and 0.0032 ± 0.0028 (ΔR2). For height, ΔR2 is 0.029 ± 0.0037. Conclusion Considering the high heritability of hypertension and height of UK Biobank and sample size of UK Biobank, sPRS results from AD databases are inflated. Independence between discovery and test sets is a well-known basic requirement for PRS studies. However, a lot of PRS studies cannot follow such requirements because of impossible direct comparisons when using summary statistics. Thus, for sPRS, potential duplications should be carefully considered within the same ethnic group
Additional file 7 of Overestimated prediction using polygenic prediction derived from summary statistics
Additional file 7: Fig. S1. The number of test set subjects required to gain statistical significance (P < 0.01) for hypertension using UK Bioban
Additional file 5 of Overestimated prediction using polygenic prediction derived from summary statistics
Additional file 5: Table S5. PRS performance comparisons for height in UK Bioban