Mamba-Chat: A Chat LLM based on State Space Models

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
  • mamba

  • You might have come across the paper Mamba paper in the last days, which was the first attempt at scaling up state space models to 2.8B parameters to work on language data.

  • mamba-chat

    Mamba-Chat: A chat LLM based on the state-space model architecture 🐍

  • Feel free to check out our Github or Huggingface repository! Our Github repo includes a cli chat script, so you can easily run the model if you have access to a GPU.

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • text-generation-webui

    A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models. (by trap20)

  • Seems to run in my hacked together text-generation-webui branch for mamba-ssm: https://github.com/trap20/text-generation-webui/tree/mamba-ssm

  • llama.cpp

    LLM inference in C/C++

  • onnxruntime

    ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator

  • causal-conv1d

    Causal depthwise conv1d in CUDA, with a PyTorch interface

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts