simple-llm-finetuner
content-chatbot
simple-llm-finetuner | content-chatbot | |
---|---|---|
12 | 5 | |
1,977 | 509 | |
- | - | |
10.0 | 6.9 | |
5 months ago | 4 months ago | |
Jupyter Notebook | Python | |
MIT License | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
simple-llm-finetuner
-
Ask HN: Resource to learn how to train and use ML Models
Just the appropriate reddit groups and follow folks on twitter, plus use a search engine.
1. Learn to run a model, checkout llama.cpp Tons of free models on huggingface.com
2. Learn to finetune a model - https://github.com/lxe/simple-llm-finetuner
3. Learn to train one. PyTorch, TensorFlow, HuggingFace libraries, etc.
Good luck.
- How can I train my custom dataset on top of Vicuna?
-
[D] The best way to train an LLM on company data
So as far as set up goes, you just need to: “”” Git clone https://github.com/lxe/simple-llama-finetuner Cd simple-llama-finetuner Pip install -r requirements.txt Python app.py ## if you’re on a remote machine (Paperspace is my go to) then you may need to edit the last line of this script to set ‘share=True’ in the launch args “””
-
Show HN: Document Q&A with GPT: web, .pdf, .docx, etc.
oobabooga's textgen webui has a tab for fine tuning now. You only need a single consumer GPU to fine tune up to 33B parameter models at a rate of about 200 epochs per hour, per GPU.
There are also one-click finetuning projects which run on free Google Colab GPUs like https://github.com/lxe/simple-llama-finetuner
It's easy and not complex at all.
-
How do I fine tune 4 bit or 8 bit models?
for a single 4090, easiest way to get started and simple to use: https://github.com/lxe/simple-llama-finetuner
- Are there publicly available datasets other than Alpaca that we can use to fine-tune LLaMA?
- Show HN: Finetune LLaMA-7B on commodity GPUs using your own text
- [Project] Finetune LLaMA-7B on commodity GPUs (and Colab) using your own text
content-chatbot
-
Show HN: Document Q&A with GPT: web, .pdf, .docx, etc.
I built this repo to do this for your own website content, that should get you a good starting point:
https://github.com/mpaepper/content-chatbot
- Your website's content -> Q&A bot / chatbot
-
Repo to create embeddings of your website's content for a Q&A bot / chatbot
Thanks for sharing the code. What happen when the existing content get updated and new contents created, would it need to create embeddings for all contents again? The current approach is not good as create embeddings cost money? Please see https://github.com/mpaepper/content-chatbot/blob/main/create.... Would it be possible progressively update the vector store?
Please advise. Thank you.
What are some alternatives?
alpaca-lora - Instruct-tune LLaMA on consumer hardware
qdrant - Qdrant - High-performance, massive-scale Vector Database for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/
paper-qa - LLM Chain for answering questions from documents with citations
faiss - A library for efficient similarity search and clustering of dense vectors.
peft - 🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Made-With-ML - Learn how to design, develop, deploy and iterate on production-grade ML applications.
slothbot - SlothBot | A generally useful analytical Discord bot that does support and writes SQL.
minimal-llama
OpenChatKit
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
WebChatRWKVstic - ChatGPT-like Web UI for RWKVstic