OpenMMLab Pre-training Toolbox and Benchmark
-
Updated
May 7, 2024 - Python
OpenMMLab Pre-training Toolbox and Benchmark
OpenMMLab Self-Supervised Learning Toolbox and Benchmark
SCAN: Learning to Classify Images without Labels, incl. SimCLR. [ECCV 2020]
solo-learn: a library of self-supervised methods for visual representation learning powered by Pytorch Lightning
PASSL包含 SimCLR,MoCo v1/v2,BYOL,CLIP,PixPro,simsiam, SwAV, BEiT,MAE 等图像自监督算法以及 Vision Transformer,DEiT,Swin Transformer,CvT,T2T-ViT,MLP-Mixer,XCiT,ConvNeXt,PVTv2 等基础视觉算法
Unsupervised Semantic Segmentation by Contrasting Object Mask Proposals. [ICCV 2021]
Unofficial implementation with pytorch DistributedDataParallel for "MoCo: Momentum Contrast for Unsupervised Visual Representation Learning"
[CVPR 2021] "The Lottery Tickets Hypothesis for Supervised and Self-supervised Pre-training in Computer Vision Models" Tianlong Chen, Jonathan Frankle, Shiyu Chang, Sijia Liu, Yang Zhang, Michael Carbin, Zhangyang Wang
Revisiting Contrastive Methods for Unsupervised Learning of Visual Representations. [NeurIPS 2021]
an implementation of MoCo and MoCo-v2 improvements pre-trained on Imagenette
A pytorch reimplement of paper "Momentum Contrast for Unsupervised Visual Representation Learning"
Self-Supervised Learning in PyTorch
MoCo v2 Pytorch tutorial, https://arxiv.org/abs/2003.04297
TF 2.x implementation of MoCo v1 (Momentum Contrast for Unsupervised Visual Representation Learning, CVPR 2020) and MoCo v2 (Improved Baselines with Momentum Contrastive Learning, 2020).
Codebase for Imperial MSc AI Group Project: How Well Do Self-Supervised Models Transfer to Medical Imaging?
Awesome-Representation-Learning-CV-PaperAndCodes, lasted development in the representation learning area.
Training MoCoV2 on the CIFAR10 Dataset
Add a description, image, and links to the moco topic page so that developers can more easily learn about it.
To associate your repository with the moco topic, visit your repo's landing page and select "manage topics."