Our great sponsors
-
smolrsrwkv
A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit evaluation. It can also directly load PyTorch RWKV models.
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
It can now quantize to 8bit for 4x memory savings: https://github.com/KerfuffleV2/smolrsrwkv/tree/experiment-quantize
This is a simple Rust implementation of RWKV. Most LLMs use Transformers (i.e. ChatGPT, etc). The creator of the RWKV approach claims it has benefits: https://github.com/BlinkDL/ChatRWKV
NOTE:
The number of mentions on this list indicates mentions on common posts plus user suggested alternatives.
Hence, a higher number means a more popular project.
Related posts
- How the RWKV language model works
- KoboldCpp - Combining all the various ggml.cpp CPU LLM inference projects with a WebUI and API (formerly llamacpp-for-kobold)
- [P] Raven 7B & 14B 🐦(RWKV finetuned on Alpaca+CodeAlpaca+Guanaco) and Gradio Demo for Raven 7B
- [D] Totally Open Alternatives to ChatGPT
- [R] RWKV 14B ctx8192 is a zero-shot instruction-follower without finetuning, 23 token/s on 3090 after latest optimization (16G VRAM is enough, and you can stream layers to save more VRAM)