Our great sponsors
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
ONNX opens an avenue for direct inference using a number of languages and platforms. For example, a model could be run directly on Android to limit data sent to a third party service. ONNX is an exciting development with a lot of promise. Microsoft has also released Hummingbird which enables exporting traditional models (sklearn, decision trees, logistical regression..) to ONNX.
The ONNX pipeline also supports exporting sentence embeddings models trained with the sentence-transformers package.
The ONNX runtime provides a common serialization format for machine learning models. ONNX supports a number of different platforms/languages and has features built in to help reduce inference time.
Related posts
- Treebomination: Convert a scikit-learn decision tree into a Keras model
- [D] GPU-enabled scikit-learn
- I learned about Microsoft's Hummingbird library today. 1000x performance??
- [D] Microsoft library, Hummingbird, compiles trained ML models into tensor computation for faster inference.
- Implementing a ChatGPT-like LLM from scratch, step by step