Our great sponsors
-
ollama-webui
Discontinued ChatGPT-Style WebUI for LLMs (Formerly Ollama WebUI) [Moved to: https://github.com/open-webui/open-webui]
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
On mac silicon:
https://ollama.ai/
ollama pull mixtral
For a chatgpt-esk web ui
https://github.com/ollama-webui/ollama-webui
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v ollama-webui:/app/backend/data --name ollama-webui --restart always ghcr.io/ollama-webui/ollama-webui:main
Navigate to http://localhost:3000
You can also use ollama in langchain.
On mac silicon:
https://ollama.ai/
ollama pull mixtral
For a chatgpt-esk web ui
https://github.com/ollama-webui/ollama-webui
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v ollama-webui:/app/backend/data --name ollama-webui --restart always ghcr.io/ollama-webui/ollama-webui:main
Navigate to http://localhost:3000
You can also use ollama in langchain.
CogVLM is very good in my (brief) testing: https://github.com/THUDM/CogVLM
The model weights seem to be under a non-commercial license, not true open source, but it is "open access" as you requested.
Lmstudio (that they linked) is definitely not open source, and doesn't even offer a pricing model for business use.
Llmstudio is, but I suspect that was a typo in their comment. https://github.com/TensorOpsAI/LLMStudio
> The output quality is not "ruined" at all.
That was my experience as well.
I also tried 2-bit version, and it was horrible.
However, there is a new approach in the works[1] (merged yesterday) which works surprisingly well with 2.10 bits per weight (12.3 GB model size)
[1] https://github.com/ggerganov/llama.cpp/pull/4773