kobold_assistant VS whisper.cpp

Compare kobold_assistant vs whisper.cpp and see what are their differences.

kobold_assistant

Like ChatGPT's voice conversations with an AI, but entirely offline/private/trade-secret-friendly, using local AI models such as LLama 2 and Whisper (by lee-b)
Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
kobold_assistant whisper.cpp
4 187
107 31,174
- -
7.1 9.8
5 months ago 1 day ago
Python C
GNU Affero General Public License v3.0 MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

kobold_assistant

Posts with mentions or reviews of kobold_assistant. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2023-07-06.
  • Is there anything to really chat with an AI as you would do with somebody over the phone ?
    4 projects | /r/LocalLLaMA | 6 Jul 2023
    My project, https://github.com/lee-b/kobold_assistant, is similar to this: currently works locally, listens and responds in a loop, but doesn't function (well) remotely through a server or anything. There is a mumble branch that works over a mumble voice chat, but it has echo/feedback/noise reduction problems right now.
  • How do I make LLMs useful to me personally?
    1 project | /r/LocalLLaMA | 26 May 2023
    KoboldAI has text completion and text chat interface built in, and I built (or am building) https://github.com/lee-b/kobold_assistant to provide a voice chat interface to that, and I have an 'ai' command that I can run from a linux shell, for any time I want the same AI to assist with anything in a command pipeline, and some steam games already use it. LocalAI will provide a local adapter to make OpenAI-based stuff use a local model instead -- I don't think this talks to KoboldAI, but I plan to modify it to do so, if it doesn't already. Finally, I plan to run (or build) something similar to SillyTavern, but instead of "joke"/fun anime characters, make it like an advisory council, where there's a philosophy advisor, a scientific advisor, an economics advisor, and so on all in a chat room, and I can just post ideas or articles I'm interested in, to get feedback from multiple perspectives, all driven by one AI -- or maybe a small number of AIs, like a specialised medical model or science or law model, and then a general model for everything else.
  • Success with a local voice chat agent
    4 projects | /r/LocalLLaMA | 26 May 2023
    kobold-assistant uses whisper. https://github.com/lee-b/kobold_assistant

whisper.cpp

Posts with mentions or reviews of whisper.cpp. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-03-31.

What are some alternatives?

When comparing kobold_assistant and whisper.cpp you can also consider the following projects:

iris-llm - IRIS: Intelligent Residential Integration System - a mind for your home!

faster-whisper - Faster Whisper transcription with CTranslate2

Whisper - High-performance GPGPU inference of OpenAI's Whisper automatic speech recognition (ASR) model

bark - 🔊 Text-Prompted Generative Audio Model

whisper - Robust Speech Recognition via Large-Scale Weak Supervision

whisperX - WhisperX: Automatic Speech Recognition with Word-level Timestamps (& Diarization)

llama.cpp - LLM inference in C/C++

NeMo - A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)

TensorRT - NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.

llama - Inference code for Llama models

text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

frogbase - Transform audio-visual content into navigable knowledge.