open-llms
WizardLM
open-llms | WizardLM | |
---|---|---|
22 | 38 | |
10,168 | 7,531 | |
- | - | |
7.7 | 9.4 | |
about 1 month ago | 7 months ago | |
Python | ||
Apache License 2.0 | - |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
open-llms
-
7 SAAS ideas 💡 you can steal
Everyone knows about ChatGPT by now, but did you know there are other models like "Mistral" or "Falcon" - you can view a full list of open-source models here or on huggingface.
- eugeneyan/open-llms
-
GPT-4 API general availability
This is the most well-maintained list of commercially usable open LLMs: https://github.com/eugeneyan/open-llms
MPT, OpenLLaMA, and Falcon are probably the most generally useful.
For code, Replit Code (specifically replit-code-instruct-glaive) and StarCoder (WizardCoder-15B) are the current top open models and both can be used commercially.
-
Local LLMs: After Novelty Wanes
There's also MPT, which has a 7B, and Falcon, with a 7B and 40B although they have not had the inference tuning in community projects that the llamas have had. This is a good repo for reviewing what's available atm: https://github.com/eugeneyan/open-llms
- How to keep track of all the LLMs out there?
-
How do I learn AI/Machine Learning?
If I was going to do the same I would at least build off of something, check out https://github.com/eugeneyan/open-llms, you should at least have a decent understanding of artificial neural networks (ANNs) and this link is pretty good on the basic concepts you need inc classification and learning types, good luck friend.
- LLM and privacy
- Local LLM to learn, explore and use for commercial purpose
-
Best instruct model recommendations to use with T4?
This list might help: https://github.com/eugeneyan/open-llms
- [D] What is the best open source LLM so far?
WizardLM
- FLaNK AI-April 22, 2024
-
Refact LLM: New 1.6B code model reaches 32% HumanEval and is SOTA for the size
This is interesting work, and a good contribution, but there is no need to mislead people.
[1] https://github.com/nlpxucan/WizardLM
-
Continue with LocalAI: An alternative to GitHub's Copilot that runs everything locally
If you pair this with the latest WizardCoder models, which have a fairly better performance than the standard Salesforce Codegen2 and Codegen2.5, you have a pretty solid alternative to GitHub Copilot that runs completely locally.
- WizardCoder context?
- The world's most-powerful AI model suddenly got 'lazier' and 'dumber.' A radical redesign of OpenAI's GPT-4 could be behind the decline in performance.
-
Official WizardLM-13B-V1.1 Released! Train with Only 1K Data! Can Achieve 86.32% on AlpacaEval!
(We will update the demo links in our github.)
-
GPT-4 API general availability
In terms of speed, we're talking about 140t/s for 7B models, and 40t/s for 33B models on a 3090/4090 now.[1] (1 token ~= 0.75 word) It's quite zippy. llama.cpp performs close on Nvidia GPUs now (but they don't have a handy chart) and you can get decent performance on 13B models on M1/M2 Macs.
You can take a look at a list of evals here: https://llm-tracker.info/books/evals/page/list-of-evals - for general usage, I think home-rolled evals like llm-jeopardy [2] and local-llm-comparison [3] by hobbyists are more useful than most of the benchmark rankings.
That being said, personally I mostly use GPT-4 for code assistance to that's what I'm most interested in, and the latest code assistants are scoring quite well: https://github.com/abacaj/code-eval - a recent replit-3b fine tune the human-eval results for open models (as a point of reference, GPT-3.5 gets 60.4 on pass@1 and 68.9 on pass@10 [4]) - I've only just started playing around with it since replit model tooling is not as good as llamas (doc here: https://llm-tracker.info/books/howto-guides/page/replit-mode...).
I'm interested in potentially applying reflexion or some of the other techniques that have been tried to even further increase coding abilities. (InterCode in particular has caught my eye https://intercode-benchmark.github.io/)
[1] https://github.com/turboderp/exllama#results-so-far
[2] https://github.com/aigoopy/llm-jeopardy
[3] https://github.com/Troyanovsky/Local-LLM-comparison/tree/mai...
[4] https://github.com/nlpxucan/WizardLM/tree/main/WizardCoder
-
WizardLM-13B-V1.0-Uncensored
You talking about this? https://github.com/nlpxucan/WizardLM
-
What 7b llm to use
The smallest model that is close to competent at code is WizardCoder 15B.. https://github.com/nlpxucan/WizardLM/
-
16-Jun-2023
WizardCoder: Empowering Code Large Language Models with Evol-Instruct (https://github.com/nlpxucan/WizardLM/tree/main/WizardCoder)
What are some alternatives?
SillyTavern - LLM Frontend for Power Users.
private-gpt - Interact with your documents using the power of GPT, 100% privately, no data leaks
SillyTavern-Extras - Extensions API for SillyTavern.
llm-humaneval-benchmarks
FastChat - An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
exllama - A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
llm-jeopardy - Automated prompting and scoring framework to evaluate LLMs using updated human knowledge prompts
airoboros - Customizable implementation of the self-instruct paper.
azure-search-openai-demo - A sample app for the Retrieval-Augmented Generation pattern running in Azure, using Azure AI Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.
promptfoo - Test your prompts, models, and RAGs. Catch regressions and improve prompt quality. LLM evals for OpenAI, Azure, Anthropic, Gemini, Mistral, Llama, Bedrock, Ollama, and other local & private models with CI/CD integration.
panml - PanML is a high level generative AI/ML development and analysis library designed for ease of use and fast experimentation.
can-ai-code - Self-evaluating interview for AI coders