3 research outputs found
Multi-view self-supervised learning for multivariate variable-channel time series
Labeling of multivariate biomedical time series data is a laborious and
expensive process. Self-supervised contrastive learning alleviates the need for
large, labeled datasets through pretraining on unlabeled data. However, for
multivariate time series data, the set of input channels often varies between
applications, and most existing work does not allow for transfer between
datasets with different sets of input channels. We propose learning one encoder
to operate on all input channels individually. We then use a message passing
neural network to extract a single representation across channels. We
demonstrate the potential of this method by pretraining our model on a dataset
with six EEG channels and then fine-tuning it on a dataset with two different
EEG channels. We compare models with and without the message passing neural
network across different contrastive loss functions. We show that our method,
combined with the TS2Vec loss, outperforms all other methods in most settings.Comment: To appear in proceedings of 2023 IEEE International workshop on
Machine Learning for Signal Processin
On convex conceptual regions in deep network representations
The current study of human-machine alignment aims at understanding the
geometry of latent spaces and the correspondence to human representations.
G\"ardenfors' conceptual spaces is a prominent framework for understanding
human representations. Convexity of object regions in conceptual spaces is
argued to promote generalizability, few-shot learning, and intersubject
alignment. Based on these insights, we investigate the notion of convexity of
concept regions in machine-learned latent spaces. We develop a set of tools for
measuring convexity in sampled data and evaluate emergent convexity in layered
representations of state-of-the-art deep networks. We show that convexity is
robust to basic re-parametrization, hence, meaningful as a quality of
machine-learned latent spaces. We find that approximate convexity is pervasive
in neural representations in multiple application domains, including models of
images, audio, human activity, text, and brain data. We measure convexity
separately for labels (i.e., targets for fine-tuning) and other concepts.
Generally, we observe that fine-tuning increases the convexity of label
regions, while for more general concepts, it depends on the alignment of the
concept with the fine-tuning objective. We find evidence that pre-training
convexity of class label regions predicts subsequent fine-tuning performance