Chinese-LLaMA-Alpaca
llm-leaderboard
Chinese-LLaMA-Alpaca | llm-leaderboard | |
---|---|---|
4 | 6 | |
17,466 | 270 | |
- | - | |
8.3 | 7.8 | |
16 days ago | 9 months ago | |
Python | Python | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Chinese-LLaMA-Alpaca
-
Chinese-Alpaca-Plus-13B-GPTQ
I'd like to share with you today the Chinese-Alpaca-Plus-13B-GPTQ model, which is the GPTQ format quantised 4bit models of Yiming Cui's Chinese-LLaMA-Alpaca 13B for GPU reference.
-
How to train a new language that is not in base model?
Could follow what people did with the Chinese-LLaMA, just for Korean. Might want to have a pure Korean corpus before feeding in a translation dataset. How big is it by the way?
- Open Source Chinese LLMs
-
Its possible to fine tune the llama model to better understand another language?
Chinese: https://github.com/ymcui/Chinese-LLaMA-Alpaca
llm-leaderboard
-
Email Obfuscation Rendered Almost Ineffective Against ChatGPT
This is assuming you’re using a really big LLM behind a paid service. There are plenty of smaller open source models. Not sure at what point it’s not “large” but when fine tuned they are capable of matching the largest LLM in performance on narrow tasks.
Some of these open source models can even be run on your local machine. It’d be very inexpensive to run thousands of pages through it.
https://llm-leaderboard.streamlit.app/
-
Is the ChatGPT and Bing AI boom already over?
palm-2-l-instruct scores 0.909 on Winogrande few-shot.
https://github.com/LudwigStumpp/llm-leaderboard/blob/main/RE...
-
Meta is preparing to launch a new open source coding model, dubbed Code Llama, that may release as soon as next week
They said it "rivals OpenAI’s Codex model" which performs worse than starcoder-16b on HumanEval-Python (pass@1) according to https://github.com/LudwigStumpp/llm-leaderboard
- All Model Leaderboards (that I know)
-
GPT-3.5 and GPT-4 performance in Open LLM Leaderboard tests?
Yes, see this leaderboard for a comparison: https://llm-leaderboard.streamlit.app/
-
Sharing my comparison methodology for LLM models
So I've tried to use a basic matrix factorization method to estimate unknown benchmark scores for models based on the known benchmark scores. Basically, I assume each model has some intrinsic "quality" score, and the known benchmarks are assumed to be a linear function of the quality score. This is similar to matrix factorization with only 1 latent factor (though the bias values have to handled differently). Then I fit the known benchmark scores from https://github.com/LudwigStumpp/llm-leaderboard to my parameters, and estimate the remaining benchmark scores.
What are some alternatives?
ChatGLM2-6B - ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
llm-humaneval-benchmarks
CodeCapybara - Open-source Self-Instruction Tuning Code LLM
chain-of-thought-hub - Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
LLMSurvey - The official GitHub page for the survey paper "A Survey of Large Language Models".
EvalAI - :cloud: :rocket: :bar_chart: :chart_with_upwards_trend: Evaluating state of the art in AI
paxml - Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
searchGPT - Grounded search engine (i.e. with source reference) based on LLM / ChatGPT / OpenAI API. It supports web search, file content search etc.
Qwen-VL - The official repo of Qwen-VL (通义千问-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.
unilm - Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
LLM-Agent-Paper-List - The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
alpa - Training and serving large-scale neural networks with auto parallelization.