What are you doing about self-hosted AI Models?

This page summarizes the projects mentioned and recommended in the original post on /r/selfhosted

InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
  • whisper.cpp

    Port of OpenAI's Whisper model in C/C++

  • First, I cloned whisper.cpp and built it with cuBLAS enabled (you might need different options depending on what hardware you are using).

  • go-subgen

    Automatically generate subtitles for your media using whisper.cpp via webhooks with support for Radarr & Sonarr

  • Have you looked at https://github.com/khakers/go-subgen to maybe automate a lot of that?

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • text-generation-webui

    A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, EXL2, llama.cpp (GGUF), Llama models.

  • You could try running one of the vicuna variants with oobagooba. I played a bit with the 13 billion parameter version on my 3060 and it was fine. I think there are more impressive models now but you should be able to run a larger model on your card.

  • automatic

    SD.Next: Advanced Implementation of Stable Diffusion and other Diffusion-based generative image models

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts