[D] Tutorial: Run LLaMA on 8gb vram on windows (thanks to bitsandbytes 8bit quantization)

This page summarizes the projects mentioned and recommended in the original post on /r/MachineLearning

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • text-generation-webui

    A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

  • conda create -n textgen conda activate textgen conda install torchvision torchaudio pytorch-cuda=11.7 git -c pytorch -c nvidia git clone https://github.com/oobabooga/text-generation-webui cd text-generation-webui pip install -r requirements.txt

  • put libbitsandbytes_cuda116.dll in C:\Users\xxx\miniconda3\envs\textgen\lib\site-packages\bitsandbytes\

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • llama-cpu

    Fork of Facebooks LLaMa model to run on CPU

  • I tried to port the llama-cpu version to a gpu-accelerated mps version for macs, it runs, but the outputs are not as good as expected and it often gives "-1" tokens. Any help and contributions on fixing it are welcome!

  • llama-mps

    Experimental fork of Facebooks LLaMa model which runs it with GPU acceleration on Apple Silicon M1/M2

  • I tried to port the llama-cpu version to a gpu-accelerated mps version for macs, it runs, but the outputs are not as good as expected and it often gives "-1" tokens. Any help and contributions on fixing it are welcome!

  • awesome-ml

    Curated list of useful LLM / Analytics / Datascience resources

  • use the prebuilt windows wheels or my WSL2 solution

  • one-click-installers

    Discontinued Simplified installers for oobabooga/text-generation-webui.

  • transformers

    🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

  • git clone https://github.com/huggingface/transformers.git cd transformers pip install -e .

  • SaaSHub

    SaaSHub - Software Alternatives and Reviews. SaaSHub helps you find the best software and product alternatives

    SaaSHub logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Schedule-Free Learning – A New Way to Train

    3 projects | news.ycombinator.com | 6 Apr 2024
  • HuggingFace Transformers: Qwen2

    1 project | news.ycombinator.com | 11 Jan 2024
  • HuggingFace Transformers Release v4.36: Mixtral, Llava/BakLlava, SeamlessM4T v2

    1 project | news.ycombinator.com | 13 Dec 2023
  • HuggingFace: Support for the Mixtral Moe

    1 project | news.ycombinator.com | 11 Dec 2023
  • Paris-Based Startup and OpenAI Competitor Mistral AI Valued at $2B

    4 projects | news.ycombinator.com | 10 Dec 2023