SaaSHub helps you find the best software and product alternatives Learn more →
Ggllm.cpp Alternatives
Similar projects and alternatives to ggllm.cpp
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.
-
-
InvokeAI
InvokeAI is a leading creative engine for Stable Diffusion models, empowering professionals, artists, and enthusiasts to generate and create visual media using the latest AI-driven technologies. The solution offers an industry leading WebUI, supports terminal use through a CLI, and serves as the foundation for multiple commercial products.
-
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
-
qdrant
Qdrant - High-performance, massive-scale Vector Database for the next generation of AI. Also available in the cloud https://cloud.qdrant.io/
-
-
exllama
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
-
-
-
curated-transformers
🤖 A PyTorch library of curated Transformer models and their composable components
-
torcheval
A library that contains a rich collection of performant PyTorch model metrics, a simple interface to create new metrics, a toolkit to facilitate metric computation in distributed training and tools for PyTorch model evaluations.
-
rust-llm-guide
Discontinued A guide to building, training and running large language models using Rust.
-
SaaSHub
SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives
ggllm.cpp discussion
ggllm.cpp reviews and mentions
-
Is there a way to use a quantized Falcon 40B with SillyTavern (on Apple Silicon)
I'd like to try https://huggingface.co/TheBloke/WizardLM-Uncensored-Falcon-40B-GGML with SillyTavern (running on Apple Silicon). The only way I've found to run Falcon 40B quantized on Apple Silicon is with https://github.com/cmp-nct/ggllm.cpp but I haven't figured out any way to get SillyTavern to use that as a local model. Does anyone know of a way to get this working?
-
How Is LLaMa.cpp Possible?
It doesn't support Falcon right now, but there's a fork that does (https://github.com/cmp-nct/ggllm.cpp/).
- Alfred-40B, an OSS RLHF version of Falcon40B
-
Falcon ggml/ggcc with langchain
To load falcon models with the new file format ggcc wich is a new file format similar to ggml, I'm using this tool: https://github.com/cmp-nct/ggllm.cpp Wich is a fork from : https://github.com/ggerganov/llama.cpp
-
Show HN: Danswer – open-source question answering across all your docs
The GGLLM fork seems to be the leading falcon winner for now [1]
It comes with its own variant of the GGML sub format "ggcv1" but there's quants available on HF [2]
Although if you have a GPU I'd go with the newly released AWQ quantization instead [3] the performance is better.
(I may or may not have a mild local LLM addiction - and video cards cost more then drugs)
[1] https://github.com/cmp-nct/ggllm.cpp
[2] https://huggingface.co/TheBloke/falcon-7b-instruct-GGML
[3] https://huggingface.co/abhinavkulkarni/tiiuae-falcon-7b-inst...
-
ChatGPT loses users for first time, shaking faith in AI revolution
For base tooling, things like:
https://huggingface.co/ (finding models and downloading them)
https://github.com/ggerganov/llama.cpp (llama)
https://github.com/cmp-nct/ggllm.cpp (falcon)
For interactive work (art/chat/research/playing around), things like:
https://github.com/oobabooga/text-generation-webui/blob/main... (llama) (Also - they just added a decent chat server built into llama.cpp the project)
https://github.com/invoke-ai/InvokeAI (stable-diffusion)
Plus a bunch of hacked together scripts.
Some example models (I'm linking to quantized versions that someone else has made, but the tooling is in the above repos to create them from the published fp16 models)
https://huggingface.co/TheBloke/llama-65B-GGML
https://huggingface.co/TheBloke/falcon-40b-instruct-GPTQ
https://huggingface.co/TheBloke/Wizard-Vicuna-30B-Uncensored...
etc. Hugging face has quite a number, although some require filling out forms for the base models for tuning/training.
- Falcon LLM – A 40B Model
-
Run machine learning on 7900XT/7900XTX using ROCm 5.5.0 on Ubuntu 22.04
I did another test running LLM model (gpt4all-falcon) quantized to Q5_0 and Q5_1 to AMD GPU (https://huggingface.co/nomic-ai/gpt4all-falcon). I used this awesome project (forked from https://github.com/ggerganov/llama.cpp to https://github.com/cmp-nct/ggllm.cpp). I hipified the CUDA file into HIP code. and made some modifications on it (PR: https://github.com/cmp-nct/ggllm.cpp/pull/3). Checkout https://huggingface.co/nomic-ai/gpt4all-falcon
-
A note from our sponsor - SaaSHub
www.saashub.com | 15 Jun 2024
Stats
cmp-nct/ggllm.cpp is an open source project licensed under GNU General Public License v3.0 or later which is an OSI approved license.
The primary programming language of ggllm.cpp is C.