Chinese-LLaMA-Alpaca
ChatGLM2-6B
Chinese-LLaMA-Alpaca | ChatGLM2-6B | |
---|---|---|
4 | 4 | |
17,539 | 15,546 | |
- | 1.1% | |
8.3 | 6.6 | |
22 days ago | about 1 month ago | |
Python | Python | |
Apache License 2.0 | GNU General Public License v3.0 or later |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
Chinese-LLaMA-Alpaca
-
Chinese-Alpaca-Plus-13B-GPTQ
I'd like to share with you today the Chinese-Alpaca-Plus-13B-GPTQ model, which is the GPTQ format quantised 4bit models of Yiming Cui's Chinese-LLaMA-Alpaca 13B for GPU reference.
-
How to train a new language that is not in base model?
Could follow what people did with the Chinese-LLaMA, just for Korean. Might want to have a pure Korean corpus before feeding in a translation dataset. How big is it by the way?
- Open Source Chinese LLMs
-
Its possible to fine tune the llama model to better understand another language?
Chinese: https://github.com/ymcui/Chinese-LLaMA-Alpaca
ChatGLM2-6B
- Are We Overlooking China's Progress in AI?
-
A new open-source language model claims to have surpassed GPT-4 right now. This needs to be fact-checked
If its benchmark results, eg on MMLU few-shot, hold and are indicative of its actual performance (which, mind you, isn't a given, for 6B nor for 130B), this 6B should be competitive with decent 30Bs. Plus natively long context and MQA. This is genuinely interesting, unlike boomer noises about CCP poisoning our checkpoints or whatever.
What are some alternatives?
CodeCapybara - Open-source Self-Instruction Tuning Code LLM
PentestGPT - A GPT-empowered penetration testing tool
LLMSurvey - The official GitHub page for the survey paper "A Survey of Large Language Models".
gpt_academic - 为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, moss等。
paxml - Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimentation and parallelization, and has demonstrated industry leading model flop utilization rates.
Qwen-VL - The official repo of Qwen-VL (通义千问-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.
inference - Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
LLM-Agent-Paper-List - The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
llm-gateway - Gateway for secure & reliable communications with OpenAI and other LLM providers
Cornucopia-LLaMA-Fin-Chinese - 聚宝盆(Cornucopia): 中文金融系列开源可商用大模型,并提供一套高效轻量化的垂直领域LLM训练框架(Pretraining、SFT、RLHF、Quantize等)
DoctorGLM - 基于ChatGLM-6B的中文问诊模型