Our great sponsors
-
WorkOS
The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.
Models larger than 2 GB are still not officially supported in ONNX. So if you want to get your large model hosted on AWS efficiently (CPU) you'll still have to do all the quantization (employing the workaround discussed in that issue) and AWS deployment yourself.
NOTE:
The number of mentions on this list indicates mentions on common posts plus user suggested alternatives.
Hence, a higher number means a more popular project.
Related posts
- Mamba-Chat: A Chat LLM based on State Space Models
- ONNX runtime: Cross-platform accelerated machine learning
- Onnx Runtime: “Cross-Platform Accelerated Machine Learning”
- How would I go about implementing machine learning in my projects from a software engineering perspective?
- You probably don't know how to do Prompt Engineering