- Shanghai
-
13:20
(UTC +08:00) - https://zhikangniu.github.io/
CV
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Recent Transformer-based CV and related works.
A paper list of some recent Transformer-based CV works.
Summary of related papers on visual attention. Related code will be released based on Jittor gradually.
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
EVA Series: Visual Representation Fantasies from BAAI
A collection of resources and papers on Vector Quantized Variational Autoencoder (VQ-VAE) and its application