zenml
onnxruntime
Our great sponsors
zenml | onnxruntime | |
---|---|---|
33 | 53 | |
3,638 | 12,583 | |
2.6% | 4.0% | |
9.8 | 10.0 | |
1 day ago | 2 days ago | |
Python | C++ | |
Apache License 2.0 | MIT License |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
zenml
- FLaNK AI - 01 April 2024
- What are some open-source ML pipeline managers that are easy to use?
-
[P] I reviewed 50+ open-source MLOps tools. Here’s the result
Currently, you can see the integrations we support here and it includes a lot of tools in your list. I also feel I agree with your categorization (it is exactly the categorization we use in our docs pretty much). Perhaps one thing missing might be feature stores but that is a minor thing in the bigger picture.
-
[P] ZenML: Build vendor-agnostic, production-ready MLOps pipelines
GitHub: https://github.com/zenml-io/zenml
- Show HN: ZenML – Portable, production-ready MLOps pipelines
-
[D] Feedback on a worked Continuous Deployment Example (CI/CD/CT)
Hey everyone! At ZenML, we released today an integration that allows users to train and deploy models from pipelines in a simple way. I wanted to ask the community here whether the example we showcased makes sense in a real-world setting:
-
How we made our integration tests delightful by optimizing our GitHub Actions workflow
As of early March 2022 this is the new CI pipeline that we use here at ZenML and the feedback from my colleagues -- fellow engineers -- has been very positive overall. I am sure there will be tweaks, changes and refactorings in the future, but for now, this feels Zen.
-
Ask HN: Who is hiring? (March 2022)
ZenML is hiring for a Design Engineer.
ZenML is an extensible, open-source MLOps framework to create production-ready machine learning pipelines. Built for data scientists, it has a simple, flexible syntax, is cloud- and tool-agnostic, and has interfaces/abstractions that are catered towards ML workflows.
We’re looking for a Design Engineer with a multi-disciplinary skill-set who can take over the look and feel of the ZenML experience. ZenML is a tool designed for developers and we want to delight them from the moment they land on our web page, to after they start using it on their machines. We would like a consistent design experience across our many touchpoints (including the [landing page](https://zenml.io), the [docs](https://docs.zenml.io), the [blog](https://blog.zenml.io), the [podcast](https://podcast.zenml.io), our social media, the product itself which is a [python package](https://github.com/zenml-io/zenml) etc).
A lot of this job is about communicating complex ideas in a beautiful way. You could be a developer or a non-coding designer, full time or part-time, employee or freelance. We are not so picky about the exact nature of this role. If you feel like you are a visually creative designer, and are willing to get stuck in the details of technical topics like MLOps, we can’t wait to work with you!
Apply here: https://zenml.notion.site/Design-Engineer-m-f-1d1a219f18a341...
-
How to improve your experimentation workflows with MLflow Tracking and ZenML
The best place to see MLflow Tracking and ZenML being used together in a simple use case is our example that showcases the integration. It builds on the quickstart example, but shows how you can add in MLflow to handle the tracking. In order to enable MLflow to track artifacts inside a particular step, all you need is to decorate the step with @enable_mlflow and then to specify what you want logged within the step. Here you can see how this is employed in a model training step that uses the autolog feature I mentioned above:
- ZenML helps data scientists work across the full stack
onnxruntime
-
AI Inference now available in Supabase Edge Functions
Embedding generation uses the ONNX runtime under the hood. This is a cross-platform inferencing library that supports multiple execution providers from CPU to specialized GPUs.
-
Deep Learning in JavaScript
tfjs is dead, looking at the commit history. The standard now is to convert PyTorch to onnx, then use onnxruntime (https://github.com/microsoft/onnxruntime/tree/main/js/web) to run the model on the browsdr.
- FLaNK Stack 05 Feb 2024
-
Vcc – The Vulkan Clang Compiler
- slang[2] has the potential, but the meta programming part is not as strong as C++, existing libraries cannot be used.
The above conclusion is drawn from my work https://github.com/microsoft/onnxruntime/tree/dev/opencl, purely nightmare to work with thoes drivers and jit compilers. Hopefully Vcc can take compute shader more seriously.
-
Oracle-samples/sd4j: Stable Diffusion pipeline in Java using ONNX Runtime
I did. It depends what you want, for an overview of how ONNX Runtime works then Microsoft have a bunch of things on https://onnxruntime.ai, but the Java content is a bit lacking on there as I've not had time to write much. Eventually I'll probably write something similar to the C# SD tutorial they have on there but for the Java API.
For writing ONNX models from Java we added an ONNX export system to Tribuo in 2022 which can be used by anything on the JVM to export ONNX models in an easier way than writing a protobuf directly. Tribuo doesn't have full coverage of the ONNX spec, but we're happy to accept PRs to expand it, otherwise it'll fill out as we need it.
- Mamba-Chat: A Chat LLM based on State Space Models
-
VectorDB: Vector Database Built by Kagi Search
What about models besides GPT? Most of the popular vector encoding models aren't using this architecture.
If you really didn't want PyTorch/Transformers, you could consider exporting your models to ONNX (https://github.com/microsoft/onnxruntime).
- ONNX runtime: Cross-platform accelerated machine learning
- Onnx Runtime: “Cross-Platform Accelerated Machine Learning”
What are some alternatives?
MLflow - Open source platform for the machine learning lifecycle
onnx - Open standard for machine learning interoperability
metaflow - :rocket: Build and manage real-life ML, AI, and data science projects with ease!
onnx-tensorrt - ONNX-TensorRT: TensorRT backend for ONNX
seldon-core - An MLOps framework to package, deploy, monitor and manage thousands of production machine learning models
onnx-simplifier - Simplify your onnx model
Poetry - Python packaging and dependency management made easy
ONNX-YOLOv7-Object-Detection - Python scripts performing object detection using the YOLOv7 model in ONNX.
pulsechain-testnet
onnx-tensorflow - Tensorflow Backend for ONNX
proposals - Temporal proposals