airoboros
WizardVicunaLM
airoboros | WizardVicunaLM | |
---|---|---|
8 | 12 | |
948 | 711 | |
- | - | |
8.7 | 6.8 | |
about 2 months ago | 11 months ago | |
Python | ||
Apache License 2.0 | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
airoboros
- TinyLlama project aims to pretrain a 1.1B Llama model on 3T tokens
- Airoboros: Customizable implementation of the self-instruct paper
-
airoboros (tool) overhaul
Just wanted to drop a note that I overhauled the airoboros tool not the models to have most of the prompts I've been using to build the datasets, plus a couple extras.
-
(2/2) May 2023
airoboros: using large language models to fine-tune large language models (https://github.com/jondurbin/airoboros)
-
Airoboros [7B/13B]
This is a fine-tuned LlaMa model, using completely synthetic training data created by https://github.com/jondurbin/airoboros
-
airobors-13b - 98% eval vs gpt-3.5-turbo
I used airoboros, a python tool I wrote, to generate the synthetic instruction response pairs, and included a jailbreak prompt to attempt to bypass OpenAI censorship. This is the only dataset used to fine-tune the model.
-
[P] airoboros 7b - instruction tuned on 100k synthetic instruction/responses
This is a 7b parameter, fine-tuned on 100k synthetic instruction/response pairs generated by gpt-3.5-turbo using my version of self-instruct airoboros
-
[P] airoboros: a rewrite of self-instruct/alpaca synthetic prompt generation
GitHub Repo
WizardVicunaLM
-
WizardLM-13B-V1.0-Uncensored
HELP! I need some clarification. I'm familiar with Wizard-Vicuna-13b-Uncensored which is EHartford's uncensoring of WizardVicunaLM.
-
Ask HN: Should I cancel my GPT-4 subscription and get Copilot instead?
> I’m also open to open source models but I hear they’re not even as good as gpt3.5.
WizardVicunaLM claims ~97% performance relative to GPT3.5: https://github.com/melodysdreamj/WizardVicunaLM
It's not particularly great at generating code, but it's uncensored and writes fantastic prose. I've been using it for the last week and I'm really satisfied with where it stands.
> It’s sad that we’re stuck in this monopoly of powerful LLMs.
Won't anyone just sponsor a few months of dedicated GPU training, finetuning and quantizing so they can be held legally accountable for it's output?
I wouldn't hold my breath.
-
Wizard-Vicuna-30B-Uncensored
Also, just noticed that you may have forgotten to update the readme, which references 13b, not 30b, thought maybe that was intentional. (If you linked directly to the Github ("WizardVicunaLM"), that would make it a bit easier for people like me to follow))
-
Where we’re at with self-hosted AI today?
There are a lot of options. Right now I'm using WizardVicunaLM to great success: https://github.com/melodysdreamj/WizardVicunaLM
It combines the uncensored WizardLM data with the Vicuna tuning to create a surprisingly high-performance model. If the chart on their GitHub page is to be believed, their model approaches GPT-3.5 performance.
-
WizardLM-30B-Uncensored
Here is the codebase and dataset for WizardVicuna https://github.com/melodysdreamj/WizardVicunaLM https://github.com/lm-sys/FastChat https://huggingface.co/datasets/RyokoAI/ShareGPT52K
- LLM that combines the principles of wizardLM and vicunaLM
-
[P] airoboros 7b - instruction tuned on 100k synthetic instruction/responses
I used the same questions from WizardVicunaLM:
- Is there a "rut" that we're in on the way to general AI?
- WizardLM-13B-Uncensored
-
Weekly Megathread
https://github.com/melodysdreamj/WizardVicunaLM - Combining WizardLM and Vicuña Principle. Made by u/Clear-Jelly2873
What are some alternatives?
WizardLM - Family of instruction-following LLMs powered by Evol-Instruct: WizardLM, WizardCoder and WizardMath
text-generation-webui - A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
TinyLlama - The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
llama.cpp - LLM inference in C/C++
datablations - Scaling Data-Constrained Language Models
koboldcpp - A simple one-file way to run various GGML and GGUF models with KoboldAI's UI
chain-of-thought-hub - Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
promptfoo - Test your prompts, models, and RAGs. Catch regressions and improve prompt quality. LLM evals for OpenAI, Azure, Anthropic, Gemini, Mistral, Llama, Bedrock, Ollama, and other local & private models with CI/CD integration.
gorilla - Gorilla: An API store for LLMs
nsfw-prompt-detection-sd - NSFW Prompt Detection for Stable Diffusion
tree-of-thoughts - Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%
shap-e - Generate 3D objects conditioned on text or images