Semi-Supervised image classification is one of the most fundamental problem in computer vision, which significantly reduces the need for human labor.
Shrinking Class Space for Enhanced Certainty in Semi-Supervised Learning
To mitigate potentially incorrect pseudo labels, recent frameworks mostly set a fixed confidence threshold to discard uncertain samples.
GitHub Link
The GitHub link is https://github.com/LiheYoung/ShrinkMatchIntroduce
The GitHub repository "LiheYoung/ShrinkMatch" presents the official PyTorch implementation of the ICCV 2023 paper titled "Shrinking Class Space for Enhanced Certainty in Semi-Supervised Learning." The paper addresses the issue of uncertainty in pseudo labels by introducing the ShrinkMatch framework. Unlike fixed confidence thresholds, ShrinkMatch identifies and removes confusion classes for the top-1 class, converting uncertain samples into certain ones. This is achieved by adapting a shrunk class space and applying consistency regularization. The method shows promising results on various datasets, including CIFAR-10, CIFAR-100, STL-10, SVHN, and ImageNet-1K, outperforming previous techniques like SimMatch and FixMatch. Detailed results and training logs are provided for reference. The codebase is available for use, and the project appreciates contributions from other related codebases. If used, the paper should be cited as specified in the citation section. To mitigate potentially incorrect pseudo labels, recent frameworks mostly set a fixed confidence threshold to discard uncertain samples.Content
This codebase provides the official PyTorch implementation of our ICCV 2023 paper: In semi-supervised learning, to mitigate potentially incorrect pseudo labels, recent frameworks mostly set a fixed confidence threshold to discard uncertain samples. This practice ensures high-quality pseudo labels, but incurs a relatively low utilization of the whole unlabeled set. Our key insight is that these uncertain samples can be turned into certain ones, as long as the confusion classes for the top-1 class are detected and removed. For each uncertain sample, our proposed ShrinkMatch adaptively seeks a shrunk class space, which merely contains the original top-1 class as well as remaining less likely classes, to satisfy the pre-defined threshold, e.g., 0.95. We then impose a consistency regularization in this shrunk space. Furthermore, considering the varied reliability among uncertain samples and the gradually improved model during training, we correspondingly design two reweighting principles for our uncertain loss. We provide all training logs. You can refer to them when reproducing. Please enter the corresponding directory according to your dataset. Many thanks to SimMatch, TorchSSL, and FixMatch-PyTorch for sharing their codebases and training logs. If you find this project useful, please consider citing:Alternatives & Similar Tools
Google Gemini, a multimodal AI by DeepMind, processes text, audio, images, and more. Gemini outperforms in AI benchmarks, is optimized for varied devices, and has been tested for safety and bias, adhering to responsible AI practices.
Video ReTalking, advanced real-world talking head video according to input audio, producing a high-quality
Then transplant it to the real world to solve complex problems
LongLLaMA is a large language model designed to handle very long text contexts, up to 256,000 tokens. It's based on OpenLLaMA and uses a technique called Focused Transformer (FoT) for training. The repository provides a smaller 3B version of LongLLaMA for free use. It can also be used as a replacement for LLaMA models with shorter contexts.
Large Language and Vision Assistant