SaaSHub helps you find the best software and product alternatives Learn more →
Video-LLaMA Alternatives
Similar projects and alternatives to Video-LLaMA
-
FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
-
LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
Otter
🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.
-
Chinese-LLaMA-Alpaca
ä¸æ–‡LLaMA&Alpaca大è¯è¨€æ¨¡åž‹+本地CPU/GPUè®ç»ƒéƒ¨ç½² (Chinese LLaMA & Alpaca LLMs)
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
Video-LLaMA reviews and mentions
- Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
-
OpenAI vs Google, Detect ChatGPT Content with 99% accuracy, Navigating AI compute costs
👀 Video-LLaMA - Empower large language models with video and audio understanding capability. (link) 🦦 Otter - Multi-modal model with improved instruction-following and in-context learning ability. 🔗 Linkly.AI - AI-powered lead analytics and management platform that helps you track, analyze, and streamline your leads in one place. 🎬 Jet Cut Ready - AI plugin for Adobe Premiere Pro that automatically removes silent parts in videos. (link) 💬 HeyGen's ChatGPT Plugin - Convert text into high-quality videos using AI text and video generation.
- Video-LLaMA: Instruction-Tuned Audio-Visual Lang Model for Video Understanding
-
Unleash the Power of Video-LLaMA: Revolutionizing Language Models with Video and Audio Understanding!
Prepare to be blown away by the cutting-edge Video-LLaMA project! We're pushing the boundaries of language models by equipping them with the remarkable ability to comprehend video and audio. Get ready for an extraordinary adventure! 🌟
-
Video-LLaMA An Instruction-tuned Audio-Visual Language Model for Video Understanding
Source Code: The codebase for pre-training and fine-tuning the Video-LLaMA model as well as the model weights are available on GitHub: https://github.com/DAMO-NLP-SG/Video-LLaMA
-
Video-ChatGPT: Redefining Interactions with Visual Data
Tons of cool stuff happening in the space, also recently saw the LLaMa-Video version of this - https://github.com/DAMO-NLP-SG/Video-LLaMA
-
Meet Video-LLaMA: A Multi-Modal Framework that Empowers Large Language Models (LLMs) with the Capability of Understanding both Visual and Auditory Content in the Video
Code: https://github.com/DAMO-NLP-SG/Video-LLaMA
-
A note from our sponsor - SaaSHub
www.saashub.com | 29 Apr 2024
Stats
DAMO-NLP-SG/Video-LLaMA is an open source project licensed under BSD 3-clause "New" or "Revised" License which is an OSI approved license.
The primary programming language of Video-LLaMA is Python.
Sponsored