[Project] MLC LLM: Universal LLM Deployment with GPU Acceleration

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • mlc-llm

    Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.

  • awesome, please share some of your experiences here https://github.com/mlc-ai/mlc-llm/issues/15 if you can, we love to see support for different hws and how well they work. We updated the latest conda so likely you can just install it again

  • web-llm

    Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.

  • It’s pretty smooth to use a ML compiler to target various GPU backends - the project was originally only for WebGPUs (https://mlc.ai/web-llm/), which is around hundreds of lines, and then it only takes tens of lines to re-target it to Vulkan, Metal and CUDA!

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • SillyTavern

    Discontinued LLM Frontend for Power Users. [Moved to: https://github.com/SillyTavern/SillyTavern] (by Cohee1207)

  • Can it be used as a server, though, through an API? We already have powerful frontends like SillyTavern, which can even run on a smartphone, so combining both would be very interesting indeed.

  • MIOpen

    AMD's Machine Intelligence Library

  • More than three months behind schedule...

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Ai on a android phone?

    2 projects | /r/LocalLLaMA | 8 Dec 2023
  • MLC vs llama.cpp

    2 projects | /r/LocalLLaMA | 7 Nov 2023
  • [Project] Scaling LLama2 70B with Multi NVIDIA and AMD GPUs under 3k budget

    1 project | /r/LocalLLaMA | 21 Oct 2023
  • Scaling LLama2-70B with Multi Nvidia/AMD GPU

    2 projects | news.ycombinator.com | 19 Oct 2023
  • ROCm Is AMD's #1 Priority, Executive Says

    5 projects | news.ycombinator.com | 26 Sep 2023