LoRA in LLaMAc++? Converting to 4bit? How to use models that are split into multiple .bin ?

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
  • alpaca_lora_4bit

  • Well llama.cpp is not python, it's cpp. Maybe with https://github.com/abetlen/llama-cpp-python or just using the huggingfasce ecosystem but it's all buggy an messy right now https://github.com/johnsmith0031/alpaca_lora_4bit so I wouldn't recommend it for someone learning python

  • llama.cpp

    LLM inference in C/C++

  • [1] Conversion to ggml: https://github.com/ggerganov/llama.cpp/blob/3265b102beb7674d010644ca2a1bd30a58f9f6b5/convert.py and [2]

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • llama-cpp-python

    Python bindings for llama.cpp

  • Well llama.cpp is not python, it's cpp. Maybe with https://github.com/abetlen/llama-cpp-python or just using the huggingfasce ecosystem but it's all buggy an messy right now https://github.com/johnsmith0031/alpaca_lora_4bit so I wouldn't recommend it for someone learning python

  • codealpaca

  • Oh, I see. That makes sense. I'm also sleep deprived over here so my reading comprehension is a bit low ;|. Well in that case check out this link: https://github.com/sahil280114/codealpaca

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts