|国家预印本平台
首页|Understanding Self-Supervised Learning via Gaussian Mixture Models

Understanding Self-Supervised Learning via Gaussian Mixture Models

Understanding Self-Supervised Learning via Gaussian Mixture Models

来源:Arxiv_logoArxiv
英文摘要

Self-supervised learning attempts to learn representations from un-labeled data; it does so via a loss function that encourages the embedding of a point to be close to that of its augmentations. This simple idea performs remarkably well, yet it is not precisely theoretically understood why this is the case. In this paper we analyze self-supervised learning in a natural context: dimensionality reduction in Gaussian Mixture Models. Crucially, we define an augmentation of a data point as being another independent draw from the same underlying mixture component. We show that vanilla contrastive learning (specifically, the InfoNCE loss) is able to find the optimal lower-dimensional subspace even when the Gaussians are not isotropic -- something that vanilla spectral techniques cannot do. We also prove a similar result for "non-contrastive" self-supervised learning (i.e., SimSiam loss). We further extend our analyses to multi-modal contrastive learning algorithms (e.g., CLIP). In this setting we show that contrastive learning learns the subset of fisher-optimal subspace, effectively filtering out all the noise from the learnt representations. Finally, we corroborate our theoretical finding through synthetic data experiments.

Parikshit Bansal、Ali Kavis、Sujay Sanghavi

信息科学、信息技术计算技术、计算机技术

Parikshit Bansal,Ali Kavis,Sujay Sanghavi.Understanding Self-Supervised Learning via Gaussian Mixture Models[EB/OL].(2024-11-05)[2025-07-22].https://arxiv.org/abs/2411.03517.点此复制

评论