ctransformers
koboldcpp
ctransformers | koboldcpp | |
---|---|---|
4 | 180 | |
1,718 | 3,951 | |
- | - | |
8.6 | 10.0 | |
4 months ago | 3 days ago | |
C | C++ | |
MIT License | GNU Affero General Public License v3.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
ctransformers
-
Refact LLM: New 1.6B code model reaches 32% HumanEval and is SOTA for the size
Does ctransformer (https://github.com/marella/ctransformers#supported-models) support running refact?
I see that model type "gpt_refact" in https://huggingface.co/smallcloudai/Refact-1_6B-fim/blob/mai...
-
How do I utilize these quantized models being uploaded?
You can also use ctransformers with the ggml models if you want to use python rather than c++.
-
Langchain and self hosted LLaMA hosted API
For ggml https://github.com/marella/ctransformers/ and https://github.com/abetlen/llama-cpp-python has a decent server. https://github.com/go-skynet/LocalAI is very active too.
- Also reconnecting with Scala. Interested in LLMs
koboldcpp
- Any Online Communities on Local/Home AI?
- Koboldcpp-1.62.1 adds support for Command-R+
- Show HN: I made an app to use local AI as daily driver
-
Easiest way to show my model to my mom?
FYI this is the easiest way to host on the horde: https://github.com/LostRuins/koboldcpp
- IT Veteran... why am I struggling with all of this?
- What do you use to run your models?
- ByteDance AI researcher suggests that open source model more powerful than Gemini to be released soon
- i need some help guys
-
[Guide] How install KoboldAI in Android via Termux (Update 04-12-2023)
For more information of Koboldcpp look this guide: https://github.com/LostRuins/koboldcpp/wiki
-
SillyTavern 1.10.10 has been released
Out of curiosity, is there a specific reason for this? The most popular fork KoboldCpp is in active development, and was the first to adopt the Min P sampler, and even distincts itself with the context shift feature. Just wondering what this means for the future. Thanks!
What are some alternatives?
llama-cpp-python - Python bindings for llama.cpp
KoboldAI
LangChain_PDFChat_Oobabooga - oobaboga -text-generation-webui implementation of wafflecomposite - langchain-ask-pdf-local
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
text-generation-inference - Large Language Model Text Generation Inference
TavernAI - Atmospheric adventure chat for AI language models (KoboldAI, NovelAI, Pygmalion, OpenAI chatgpt, gpt-4)
artificial-nose - Instructions, source code, and misc. resources needed for building a Tiny ML-powered artificial nose.
KoboldAI - KoboldAI is generative AI software optimized for fictional use, but capable of much more!
kendryte-standalone-sdk - Standalone SDK for kendryte K210
ChatRWKV - ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
SillyTavern - LLM Frontend for Power Users. [Moved to: https://github.com/SillyTavern/SillyTavern]