sliced_llama

Simple LLM inference server (by silphendio)

Sliced_llama Alternatives

Similar projects and alternatives to sliced_llama based on common topics and language

NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a better sliced_llama alternative or higher similarity.

sliced_llama reviews and mentions

Posts with mentions or reviews of sliced_llama. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-08.
  • Mixture-of-Depths: Dynamically allocating compute in transformers
    3 projects | news.ycombinator.com | 8 Apr 2024
    There are already some implementations out there which attempt to accomplish this!

    Here's an example: https://github.com/silphendio/sliced_llama

    A gist pertaining to said example: https://gist.github.com/silphendio/535cd9c1821aa1290aa10d587...

    Here's a discussion about integrating this capability with ExLlama: https://github.com/turboderp/exllamav2/pull/275

    And same as above but for llama.cpp: https://github.com/ggerganov/llama.cpp/issues/4718#issuecomm...

Stats

Basic sliced_llama repo stats
1
15
7.6
3 months ago

silphendio/sliced_llama is an open source project licensed under MIT License which is an OSI approved license.

The primary programming language of sliced_llama is Python.


Sponsored
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com