May. 2nd, 2024: Vision Mamba (Vim) is accepted by ICML2024. 🎉 Conference page can be found here. Feb. 10th, 2024: We update Vim-tiny/small weights and training scripts. By placing the class token at ...
Abstract: Vision-Language Models (VLMs), such as CLIP, excel in zero-shot image-level visual understanding but struggle with object-based tasks requiring precise localization and recognition. Visual ...
If you find any work missing or have any suggestions (papers, implementations, and other resources), feel free to pull requests. We will add the missing papers to this repo as soon as possible. You ...
Abstract: Learning a discriminative model to distinguish a target from its surrounding distractors is essential to generic visual object tracking. Dynamic target representation adaptation against ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results