Sources:
- For MoCo:
- MoCo v1 2019 paper
- MoCo v2 2020 paper
- Paper explained — Momentum Contrast for Unsupervised Visual Representation Learning [MoCo] by Nazim Bendib
- For BYOL:
- BYOL 2020 paper
- Neural Networks Intuitions: 10. BYOL- Paper Explanation by Raghul Asokan
- Understanding self-supervised and contrastive learning with "Bootstrap Your Own Latent" (BYOL) by imbue
MoCo and BYOL are both famous contrastive learning and self-supervised learning frameworks. They introduce an interesting designs:
- Using a online network and a fixed target network to mitigate the collapse problem in contrastive learning.
- Use the same architecture for these two networks while only training the online network. The target network is updated via EMA (expenential moving average).
This design is heavily used in further research like DINO as is worth learning about.
You may need to read my post for SimCLR to get a deeper understanding of contrastive learning.