"Samantha-33B-SuperHOT-8K-GPTQ" now that's a great name for a true model.

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
  • alpaca_lora_4bit

  • I would also like to know how one would finetune this in 4 bit? I think one could take the merged 8K PEFT with the LLaMA weights, and then quantize it to 4 bit, and then train with https://github.com/johnsmith0031/alpaca_lora_4bit ?

  • dockerLLM

    TheBloke's Dockerfiles

  • The one thing I have published is my Docker files for producing my two Runpod templates, which let people try GGML and GPTQ models on Runpod pods with full GPU acceleration (ExLlama and AutoGPTQ). They can be found at https://github.com/TheBlokeAI/dockerLLM/ .

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts