koboldcpp
runpodctl
koboldcpp | runpodctl | |
---|---|---|
180 | 8 | |
3,817 | 215 | |
- | 4.2% | |
10.0 | 9.4 | |
2 days ago | 20 days ago | |
C++ | Go | |
GNU Affero General Public License v3.0 | GNU General Public License v3.0 only |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
koboldcpp
- Any Online Communities on Local/Home AI?
- Koboldcpp-1.62.1 adds support for Command-R+
- Show HN: I made an app to use local AI as daily driver
-
Easiest way to show my model to my mom?
FYI this is the easiest way to host on the horde: https://github.com/LostRuins/koboldcpp
- IT Veteran... why am I struggling with all of this?
- What do you use to run your models?
- ByteDance AI researcher suggests that open source model more powerful than Gemini to be released soon
- i need some help guys
-
[Guide] How install KoboldAI in Android via Termux (Update 04-12-2023)
For more information of Koboldcpp look this guide: https://github.com/LostRuins/koboldcpp/wiki
-
SillyTavern 1.10.10 has been released
Out of curiosity, is there a specific reason for this? The most popular fork KoboldCpp is in active development, and was the first to adopt the Min P sampler, and even distincts itself with the context shift feature. Just wondering what this means for the future. Thanks!
runpodctl
- Ask HN: What's the best hardware to run small/medium models locally?
-
Old Timer needs help setting up stable diffusion. Extremely confused.
You can rent a GPU on https://www.runpod.io/, which also has stable diffusion templates so any time you start the GPU SD will be preinstalled for you, ready to use:)
-
i need some help guys
Another option is to use a service like www.runpod.io to rent time on more powerful systems. A few times a week I’ll load up whatever the latest 13B or 20/24B (and even low bpw 70Bs with EXL2) model is on a system with an rtx 3090 for $0.44/hr, and sometimes I’ll treat myself to an A6000 system to run 4bit 70Bs for $0.79/hr. They also offer A4000 systems with 16GB VRAM which is plenty to run a 4bpw 13B EXL2 model or an 8bpw 7B model, and those systems are just $0.36/hr.
- GPT-3.5 Turbo fine-tuning and API updates
-
What's the best (and cheap) way to try out all the new LLMs on cloud services.
Many people use sites like runpod for this.
-
Looking for Paperspace (or equivalent) Help
You can rent time on systems on www.runpod.io with a 48GB A6000 for $0.50/hr spot pricing and $0.79/hr regular pricing. 3090s can be had for $0.29/hr spot pricing and $.44/hr regular.
-
only seeing disk cache slider, no gpu anything?
Keep in mind you can also use www.runpod.io to rent access to systems with a 3090 for about $.45/hr. It might be significantly cheaper ir at least more affordable to do this for a few hours a week instead of dropping $1,000 on a new laptop. This is what I personally do (I generally use it in the evening, and can get an Nvidia A6000 with 48GB VRAM for $.49/hr spot pricing). This lets me play with the latest 33B and 65B models, with really fast replies, and I spend maybe $5-7 a week if I use it a lot.
-
What is best bang for buck persistent virtual GPU rental to run SD?
you can also connect the volume to other cloud storage (pcloud, dropbox, etc) (look for the option cloud sync in the "my pods" console) , or use python libraries to pull/push files from S3 buckets / dropbox / ftps / your NAS / etc, or use their docker tools to transfer files between your local PC and the volumehttps://github.com/runpod/runpodctl/blob/main/README.md
What are some alternatives?
KoboldAI
chroma - the AI-native open-source embedding database
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
KoboldAI-Runpod - This is just a simple set of notebooks to load koboldAI and SillyTavern Extras on a runpod with Pytorch 2.0.1 Template
TavernAI - Atmospheric adventure chat for AI language models (KoboldAI, NovelAI, Pygmalion, OpenAI chatgpt, gpt-4)
KoboldAI
EveryDream2trainer
ChatRWKV - ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
SillyTavern - LLM Frontend for Power Users. [Moved to: https://github.com/SillyTavern/SillyTavern]
exllama - A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
TavernAI - TavernAI for nerds [Moved to: https://github.com/Cohee1207/SillyTavern]