llamafile VS ok-robot

Compare llamafile vs ok-robot and see what are their differences.

llamafile

Distribute and run LLMs with a single file. (by Mozilla-Ocho)

ok-robot

An open, modular framework for zero-shot, language conditioned pick-and-drop tasks in arbitrary homes. (by ok-robot)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
llamafile ok-robot
36 7
15,120 361
29.1% -
9.6 9.5
3 days ago 2 months ago
C++ Python
GNU General Public License v3.0 or later MIT License
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

llamafile

Posts with mentions or reviews of llamafile. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-05-06.
  • FLaNK-AIM Weekly 06 May 2024
    45 projects | dev.to | 6 May 2024
  • llamafile v0.8
    1 project | news.ycombinator.com | 24 Apr 2024
  • Mistral AI Launches New 8x22B Moe Model
    4 projects | news.ycombinator.com | 9 Apr 2024
    I think the llamafile[0] system works the best. Binary works on the command line or launches a mini webserver. Llamafile offers builds of Mixtral-8x7B-Instruct, so presumably they may package this one up as well (potentially a quantized format).

    You would have to confirm with someone deeper in the ecosystem, but I think you should be able to run this new model as is against a llamafile?

    [0] https://github.com/Mozilla-Ocho/llamafile

  • Apple Explores Home Robotics as Potential 'Next Big Thing'
    3 projects | news.ycombinator.com | 4 Apr 2024
    Thermostats: https://www.sinopetech.com/en/products/thermostat/

    I haven't tried running a local text-to-speech engine backed by an LLM to control Home Assistant. Maybe someone is working on this already?

    TTS: https://github.com/SYSTRAN/faster-whisper

    LLM: https://github.com/Mozilla-Ocho/llamafile/releases

    LLM: https://huggingface.co/TheBloke/Nous-Hermes-2-Mixtral-8x7B-D...

    It would take some tweaking to get the voice commands working correctly.

  • LLaMA Now Goes Faster on CPUs
    16 projects | news.ycombinator.com | 31 Mar 2024
    While I did not succeed in making the matmul code from https://github.com/Mozilla-Ocho/llamafile/blob/main/llamafil... work in isolation, I compared eigen, openblas, and mkl: https://gist.github.com/Dobiasd/e664c681c4a7933ef5d2df7caa87...

    In this (very primitive!) benchmark, MKL was a bit better than eigen (~10%) on my machine (i5-6600).

    Since the article https://justine.lol/matmul/ compared the new kernels with MLK, we can (by transitivity) compare the new kernels with Eigen this way, at least very roughly for this one use-case.

  • Llamafile 0.7 Brings AVX-512 Support: 10x Faster Prompt Eval Times for AMD Zen 4
    3 projects | news.ycombinator.com | 31 Mar 2024
    Yes, they're just ZIP files that also happen to be actually portable executables.

    https://github.com/Mozilla-Ocho/llamafile?tab=readme-ov-file...

  • Show HN: I made an app to use local AI as daily driver
    31 projects | news.ycombinator.com | 27 Feb 2024
    have you seen llamafile[0]?

    [0] https://github.com/Mozilla-Ocho/llamafile

  • FLaNK Stack 26 February 2024
    50 projects | dev.to | 26 Feb 2024
  • Gemma.cpp: lightweight, standalone C++ inference engine for Gemma models
    7 projects | news.ycombinator.com | 23 Feb 2024
    llama.cpp has integrated gemma support. So you can use llamafile for this. It is a standalone executable that is portable across most popular OSes.

    https://github.com/Mozilla-Ocho/llamafile/releases

    So, download the executable from the releases page under assets. You want either just main or just server. Don't get the huge ones with the model inlined in the file. The executable is about 30MB in size,

    https://github.com/Mozilla-Ocho/llamafile/releases/download/...

  • Ollama releases OpenAI API compatibility
    12 projects | news.ycombinator.com | 8 Feb 2024
    The improvements in ease of use for locally hosting LLMs over the last few months have been amazing. I was ranting about how easy https://github.com/Mozilla-Ocho/llamafile is just a few hours ago [1]. Now I'm torn as to which one to use :)

    1: Quite literally hours ago: https://euri.ca/blog/2024-llm-self-hosting-is-easy-now/

ok-robot

Posts with mentions or reviews of ok-robot. We have used some of these posts to build our list of alternatives and similar projects. The last one was on 2024-04-04.
  • Apple Explores Home Robotics as Potential 'Next Big Thing'
    3 projects | news.ycombinator.com | 4 Apr 2024
  • Low Cost Robot Arm
    8 projects | news.ycombinator.com | 1 Apr 2024
    That's it, isn't it. The question is not, how far away from that are we, but when can you and I actually afford it? Because, as the other commenter snarkily replies, human maid's already exist. The lifestyle of the singularity is already here for the rich. It's trickling down that kind of lifestyle to the rest of us that AI robots will enable. (with some amount of social upheaval.)

    Lets say the robot that can do that comes out next year for $15 million. Could you afford one? I certainly can't. So pretend that it does, what changes for you and I? Nothing. So the robots that can do that won't be used as robot maids until the price comes down. Which; it will. Open source robotics and model-available AI will force things to be affordable sooner, rather than later, because we'd all like a robot to do that for us.

    The industrial versions will be used to do hideously dangerous things. underwater welding, chainsaw helicoptering, manual nuclear reactor rod removal. We already use machines for a lot of those difficult/impossible tasks, it's just a matter of programming the robots.

    Which takes us back to today. How far away from that are we? The pieces are already here. Between https://ok-robot.github.io/ and https://mobile-aloha.github.io/ the building blocks are here. It's just a matter of time before someone puts the existing pieces together to make said robot, the only question is who will be first to make it, who will be first to open source it. Who will make it not just possible, but affordable?

  • GPT-4, without specialized training, beat a GPT-3.5 class model that cost $10B
    3 projects | news.ycombinator.com | 24 Mar 2024
    Thanks! Appreciate the kind words. I should have in the next month or so (interviewing and finishing my Master's, so there's been delays) a follow up that follows more advancements in the router style VLA, sensoiromotor VLM, and advances in embedding enriched vision models in general.

    If you want a great overview of what a modern robotics stack would look like with all this, https://ok-robot.github.io/ was really good and will likely make it into the article. It's a VLA combined with existing RL methods to demonstrate multi-tasking robots, and serves as a great glimpes into what a lot of researchers are working on. You won't see these techniques in robots in industrial or commercial settings - we're still too new at this to be reliable or capable enough to deploy these on real tasks.

  • Figure robotics demos its OpenAI integration
    1 project | news.ycombinator.com | 13 Mar 2024
    The Ok-robot demo shows that the technology for it to be fairly general is there, though no idea if figure one is using their technology or not. Simply being able to command a robot instead of moving a turtle with gcode is nothing short of astounding to those who aren’t deeply involved and tracking the sota progress in this area.

    https://ok-robot.github.io/

  • FLaNK Stack 26 February 2024
    50 projects | dev.to | 26 Feb 2024
  • Show HN: OK-Robot: open, modular home robot framework for pick-and-drop anywhere
    5 projects | news.ycombinator.com | 23 Feb 2024
    Disclaimer: I'm not one of the authors, but I work in this area.

    You basically hit the nail on the head with these questions. This work is super cool, but you named a lot of the limitations with contemporary robot learning systems.

    1. It's using an object classifier. It's described here (https://github.com/ok-robot/ok-robot/tree/main/ok-robot-navi...), but if I understanding it correctly basically they are using a ViT model (basically an image classification model) to do some labeling of images and projecting them onto a voxel grid. Then they are using language embeddings from CLIP to pair the language with the voxel grid. The limitations of this are that if they want this to run on the robot, they can't use the super huge versions of these models. While they could use a huge model on the cloud, that would introduce a lot of latency.

What are some alternatives?

When comparing llamafile and ok-robot you can also consider the following projects:

ollama - Get up and running with Llama 3, Mistral, Gemma, and other large language models.

ollama-webui - ChatGPT-Style WebUI for LLMs (Formerly Ollama WebUI) [Moved to: https://github.com/open-webui/open-webui]

langchain - 🦜🔗 Build context-aware reasoning applications

LLaVA - [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.

llama.cpp - LLM inference in C/C++

safetensors - Simple, safe way to store and distribute tensors

LocalAIVoiceChat - Local AI talk with a custom voice based on Zephyr 7B model. Uses RealtimeSTT with faster_whisper for transcription and RealtimeTTS with Coqui XTTS for synthesis.

chatgpt-web - ChatGPT web interface using the OpenAI API

TinyLlama - The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.

llamafile-docker - Simple llamafile setup with docker

gemma.cpp - lightweight, standalone C++ inference engine for Google's Gemma models.

Cgml - GPU-targeted vendor-agnostic AI library for Windows, and Mistral model implementation.