OASIS
VLDet
OASIS | VLDet | |
---|---|---|
1 | 1 | |
317 | 169 | |
2.5% | - | |
10.0 | 3.1 | |
over 1 year ago | about 1 month ago | |
Python | Python | |
GNU Affero General Public License v3.0 | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
OASIS
-
StyleGAN-NADA: Blind Training and Other Wonders
Conclusion So this is how StyleGAN-NADA, a CLIP-guided zero-shot method for Non-Adversarial Domain Adaptation of image generators, works. Although the StyleGAN-NADA is focused on StyleGAN, it can be applied to other generative architectures such as OASIS and many others.
VLDet
-
[R] [ICLR'2023🌟]: Vision-and-Language Framework for Open-Vocabulary Object Detection
We're excited to share our latest work "Learning Object-Language Alignments for Open-Vocabulary Object Detection", which got accepted to ICLR'2023. Here're some resources: arxiv paper: https://arxiv.org/abs/2211.14843 github: https://github.com/clin1223/VLDet The proposed method called **VLDet**, which is a a simple yet effective vision-and-language framework for open-vocabulary object detection. Our key efforts are: 🔥 We introduce an open-vocabulary object detector method to learn object-language alignments directly from image-text pair data. 🔥 We propose to formulate region-word alignments as a set-matching problem and solve it efficiently with the Hungarian algorithm. 🔥 We use all nouns from image-text pairs as our object voccabulary which is strictly following the open-vocabulary setting and extensive experiments on two benchmark datasets, COCO and LVIS, demonstrate our superior performance.
What are some alternatives?
Deep-Exemplar-based-Video-Colorization - The source code of CVPR 2019 paper "Deep Exemplar-based Video Colorization".
CLIP-Caption-Reward - PyTorch code for "Fine-grained Image Captioning with CLIP Reward" (Findings of NAACL 2022)
Keras-GAN - Keras implementations of Generative Adversarial Networks.
DeepKE - [EMNLP 2022] An Open Toolkit for Knowledge Graph Extraction and Construction
pytorch-CycleGAN-and-pix2pix - Image-to-Image Translation in PyTorch
robo-vln - Pytorch code for ICRA'21 paper: "Hierarchical Cross-Modal Agent for Robotics Vision-and-Language Navigation"
AdamP - AdamP: Slowing Down the Slowdown for Momentum Optimizers on Scale-invariant Weights (ICLR 2021)
yolov5 - YOLOv5 🚀 in PyTorch > ONNX > CoreML > TFLite
pix2pixHD - Synthesizing and manipulating 2048x1024 images with conditional GANs
VL_adapter - PyTorch code for "VL-Adapter: Parameter-Efficient Transfer Learning for Vision-and-Language Tasks" (CVPR2022)
clean-fid - PyTorch - FID calculation with proper image resizing and quantization steps [CVPR 2022]
mmdetection - OpenMMLab Detection Toolbox and Benchmark