new serverless templates on banana for community use covering the entire code llama family

This page summarizes the projects mentioned and recommended in the original post on /r/LocalLLaMA

Scout Monitoring - Free Django app performance insights with Scout Monitoring
Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.
www.scoutapm.com
featured
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
  • demo-codellama-7b-gptq

    Discontinued This is a CodeLlama-7B-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-7b

  • Scout Monitoring

    Free Django app performance insights with Scout Monitoring. Get Scout setup in minutes, and let us sweat the small stuff. A couple lines in settings.py is all you need to start monitoring your apps. Sign up for our free tier today.

    Scout Monitoring logo
  • demo-codellama-13b-gptq

    This is a CodeLlama-13B-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-13b

  • demo-codellama-34b-gptq

    Discontinued This is a CodeLlama-34B-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-34b

  • demo-codellama-7b-python-gptq

    This is a CodeLlama-7B-Python-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-python-7b

  • demo-codellama-13b-python-gptq

    Discontinued This is a CodeLlama-13B-Python-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-python-13b

  • demo-codellama-34b-python-gptq

    This is a CodeLlama-34B-Python-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-python-34b

  • demo-codellama-7b-instruct-gptq

    Discontinued This is a CodeLlama-7B-Instruct-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-instruct-7b

  • InfluxDB

    Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.

    InfluxDB logo
  • demo-codellama-13b-instruct-gptq

    This is a CodeLlama-13B-Instruct-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-instruct-13b

  • demo-codellama-34b-instruct-gptq

    This is a CodeLlama-34B-Instruct-GPTQ model starter template from Banana.dev that allows on-demand serverless GPU inference.

  • code llama-instruct-34B

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts

  • Highly realistic talking head video generation

    1 project | news.ycombinator.com | 15 Jun 2024
  • Show HN: Apache Echarts for Jupyter

    1 project | news.ycombinator.com | 15 Jun 2024
  • Show HN: Python) Markdown Exec, execute code blocks and render their output

    1 project | news.ycombinator.com | 15 Jun 2024
  • Mathematics secret behind AI on Digit Recognition

    3 projects | dev.to | 15 Jun 2024
  • Perplexity AI Is Lying about Their User Agent

    1 project | news.ycombinator.com | 15 Jun 2024