chen20_simpl_framew_contr_learn_visual_repres: A simple framework for contrastive learning of visual representations
SimCLR is a simple framework for Contrastive Methods of visual representations.
A simple framework for contrastive learning of visual representations
We do not train the model with a memory bank
Rather than train with a memory bank, they use a large batch size, and the LARS Optimizer to stabilize training.
Key Contributions
- Composition of data augmentation to form positive pairs
- introduce a learnable non-linear transformation between the representation and the contrastive loss substantially improves the quality of the learned representations
- Contrastive learning benefits from larger batch sizes and more training steps compared to supervised learning
Data Augmentation
A stochastic data augmentation module is introduced to produce two
correlated views of the same example, denoted
- random cropping
- random color distortions
- random Gaussian blur
A neural network encoder
A small network projection head
The loss function (normalized temperature-scaled cross entropy loss)
is applied on the output of
A minibatch of N examples is sampled, resulting in
The Importance of the Projection Head
It is conjectured that the projection head
<biblio.bib>