ort VS firecracker

Compare ort vs firecracker and see what are their differences.

Our great sponsors
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • WorkOS - The modern identity platform for B2B SaaS
  • SaaSHub - Software Alternatives and Reviews
ort firecracker
7 75
555 24,084
14.6% 2.0%
9.3 9.9
9 days ago 3 days ago
Rust Rust
Apache License 2.0 Apache License 2.0
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

ort

Posts with mentions or reviews of ort. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-16.
  • AI Inference now available in Supabase Edge Functions
    4 projects | dev.to | 16 Apr 2024
    To solve this, we built a native extension in Edge Runtime that enables using ONNX runtime via the Rust interface. This was made possible thanks to an excellent Rust wrapper called Ort:
  • AI Inference Now Available in Supabase Edge Functions
    1 project | news.ycombinator.com | 16 Apr 2024
    hey hn, supabase ceo here

    As the post points out, this comes in 2 parts:

    1. Embeddings models for RAG workloads (specifically pgvector). Available today.

    2. Large Language Models for GenAI workloads. This will be progressively rolled out as we get our hands on more GPUs.

    We've always had a focus on architectures that can run anywhere (especially important for local dev and self-hosting). In that light, we've found that the Ollama[0] tooling is really unbeatable. I heard one of our engineers explain it like "docker for models" which I think is apt.

    To support models that work best with GPUs, we're running them with Fly GPUs - pretty much this: https://fly.io/blog/scaling-llm-ollama (and then we stitch a native API around it). The plan is that you will be able to "BYO" model server and point the Edge Runtime towards it using simple env vars / config.

    We've also made improvements for CPU models. We built a native extension in Edge Runtime that enables using ONNX runtime via the Rust interface. This was made possible thanks to an excellent Rust wrapper, Ort[1]. We have the models stored on disk, so there is no downloading, cold-boot, etc.

    The thing I most like about this set up is that you can now use Edge Functions like background workers for your Postgres database, offloading heavy compute for generating embeddings. For example, you can trigger the worker when a user inserts some text, and then the worker will asynchronously create the embedding and store it back into your database.

    I'll be around if there are any questions.

    [0] ollama.com

    [1] Ort: https://github.com/pykeio/ort

  • Moving from Typescript and Langchain to Rust and Loops
    9 projects | dev.to | 7 Sep 2023
    In the quest for more efficient solutions, the ONNX runtime emerged as a beacon of performance. The decision to transition from Typescript to Rust was an unconventional yet pivotal one. Driven by Rust's robust parallel processing capabilities using Rayon and seamless integration with ONNX through the ort crate, Repo-Query unlocked a realm of unparalleled efficiency. The result? A transformation from sluggish processing to, I have to say it, blazing-fast performance.
  • How to create YOLOv8-based object detection web service using Python, Julia, Node.js, JavaScript, Go and Rust
    19 projects | dev.to | 13 May 2023
    ort - ONNX runtime library.
  • Do you use Rust in your professional career?
    6 projects | /r/rust | 9 May 2023
    Our main model in Rust is a deep neural network, using ONNX via the ort rust bindings. The application is some particular applications of process automation.
  • onnxruntime
    4 projects | /r/rust | 22 Feb 2023
    You could try ort https://github.com/pykeio/ort It looks like it's in active development and supports GPU inference
  • Deep Learning in Rust: Burn 0.4.0 released and plans for 2023
    6 projects | /r/rust | 2 Jan 2023
    I would't try to distribute your ml models with the typical frameworks, especially not with python. Have you looked in to ONNX?For example: https://github.com/pykeio/ort

firecracker

Posts with mentions or reviews of firecracker. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-12.
  • Lambda Internals: Why AWS Lambda Will Not Help With Machine Learning
    1 project | dev.to | 25 Apr 2024
    This architecture leverages microVMs for rapid scaling and high-density workloads. But does it work for GPU? The answer is no. You can look at the old 2019 GitHub issue and the comments to it to get the bigger picture of why it is so.
  • Show HN: Add AI code interpreter to any LLM via SDK
    5 projects | news.ycombinator.com | 12 Apr 2024
    Hi, I'm the CEO of the company that built this SDK.

    We're a company called E2B [0]. We're building and open-source [1] secure environments for running untrusted AI-generated code and AI agents. We call these environments sandboxes and they are built on top of micro VM called Firecracker [2].

    You can think of us as giving small cloud computers to LLMs.

    We recently created a dedicated SDK for building custom code interpreters in Python or JS/TS. We saw this need after a lot of our users have been adding code execution capabilities to their AI apps with our core SDK [3]. These use cases were often centered around AI data analysis so code interpreter-like behavior made sense

    The way our code interpret SDK works is by spawning an E2B sandbox with Jupyter Server. We then communicate with this Jupyter server through Jupyter Kernel messaging protocol [4].

    We don't do any wrapping around LLM, any prompting, or any agent-like framework. We leave all of that on users. We're really just a boring code execution layer that sats at the bottom that we're building specifically for the future software that will be building another software. We work with any LLM. Here's how we added code interpreter to Claude [5].

    Our long-term plan is to build an automated AWS for AI apps and agents.

    Happy to answer any questions and hear feedback!

    [0] https://e2b.dev/

    [1] https://github.com/e2b-dev

    [2] https://github.com/firecracker-microvm/firecracker

    [3] https://e2b.dev/docs

    [4] https://jupyter-client.readthedocs.io/en/latest/messaging.ht...

    [5] https://github.com/e2b-dev/e2b-cookbook/blob/main/examples/c...

  • Fly.it Has GPUs Now
    5 projects | news.ycombinator.com | 13 Feb 2024
    As far as I know, Fly uses Firecracker for their VMs. I've been following Firecracker for a while now (even using it in a project), and they don't support GPUs out of the box (and have no plan to support it [1]).

    I'm curious to know how Fly figured their own GPU support with Firecracker. In the past they had some very detailed technical posts on how they achieved certain things, so I'm hoping we'll see one on their GPU support in the future!

    [1]: https://github.com/firecracker-microvm/firecracker/issues/11...

  • MotorOS: a Rust-first operating system for x64 VMs
    7 projects | news.ycombinator.com | 7 Jan 2024
    I pass through a GPU and USB hub to a VM running on a machine in the garage. An optical video cable and network compatible USB extender brings the interface to a different room making it my primary “desktop” computer (and an outdated laptop as a backup device). Doesn’t get more silent and cool than this. Another VM on the garage machine gets a bunch of hard drives passed through to it.

    That said, hardware passthrough/VFIO is likely out of the current realistic scope for this project. VM boot times can be optimized if you never look for hardware to initialize in the first place. Though they are still likely initializing a network interface of some sort.

    “MicroVM” seems to be a term used when as much as possible is stripped from a VM, such as with https://github.com/firecracker-microvm/firecracker

  • Virtual Machine as a Core Android Primitive
    2 projects | news.ycombinator.com | 5 Dec 2023
    According to their own FAQ it is indeed: https://github.com/firecracker-microvm/firecracker/blob/main...
  • Sandboxing a .NET Script
    1 project | /r/dotnet | 22 Oct 2023
    What about microVMs like firecracker?
  • We Replaced Firecracker with QEMU
    5 projects | news.ycombinator.com | 10 Jul 2023
    Dynamic memory management - Firecracker's RAM footprint starts low, but once a workload inside allocates RAM, Firecracker will never return it to the host system. After running several workloads inside, you end up with an idling VM that consumes 32 GB of RAM on the host, even though it doesn't need any of it.

    Firecracker has a balloon device you can inflate (ie: acquire as much memory inside the VM as possible) and then deflate... returning the memory to the host.

    https://github.com/firecracker-microvm/firecracker/blob/main...

  • I'm looking for a virtual machine that prioritizes privacy and does not include tracking or telemetry.
    1 project | /r/privacy | 5 Jun 2023
  • Neverflow: Set of C macros that guard against buffer overflows
    4 projects | news.ycombinator.com | 2 Jun 2023
    Very few things in those companies are being written in Rust, and half of those projects chose Rust around ideological reasons rather than technical, with plenty of 'unsafe' thrown in for performance reasons

    https://github.com/firecracker-microvm/firecracker/search?q=...

    The fact that 'unsafe' even exists in Rust means it's no better than C with some macros.

    Don't get me wrong, Rust has it's place, like all the other languages that came about for various reasons, but it's not going to gain wide adoption.

    Future of programming consists of 2 languages - something like C that has a small instruction set for adopting to new hardware, and something that is very high level, higher than Python with LLM in the background. Everything in the middle is fodder.

  • Do you use Rust in your professional career?
    6 projects | /r/rust | 9 May 2023
    https://github.com/firecracker-microvm/firecracker is the one that comes to mind, but most of these are internal.

What are some alternatives?

When comparing ort and firecracker you can also consider the following projects:

onnxruntime-rs - Rust wrapper for Microsoft's ONNX Runtime (version 1.8)

cloud-hypervisor - A Virtual Machine Monitor for modern Cloud workloads. Features include CPU, memory and device hotplug, support for running Windows and Linux guests, device offload with vhost-user and a minimal compact footprint. Written in Rust with a strong focus on security.

yolov8_onnx_go - YOLOv8 Inference using Go

bottlerocket - An operating system designed for hosting containers

onnxruntime-php - Run ONNX models in PHP

gvisor - Application Kernel for Containers

yolov8_onnx_javascript - YOLOv8 inference using Javascript

libkrun - A dynamic library providing Virtualization-based process isolation capabilities

langchainjs - 🦜🔗 Build context-aware reasoning applications 🦜🔗

krunvm - Create microVMs from OCI images

yolov8_onnx_julia - YOLOv8 inference using Julia

deno - A modern runtime for JavaScript and TypeScript.