-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
-
XNNPACK
High-efficiency floating-point neural network inference operators for mobile, server, and Web
There's always been a tradeoff in writing code between developer experience and taking full advantage of what the hardware is capable of. That "waste" in execution efficiency is often worth it for the sake of representing helpful abstractions and generally helping developer productivity.
The GFLOP/s is 1/28th of what you'd get when using the native Accelerate framework on M1 Macs [1]. I am all in for powerful abstractions, but not using native APIs for this (even if it's just the browser calling Accelerate in some way) is just a huge waste of everyone's CPU cycles and electricity.
[1] https://github.com/danieldk/gemm-benchmark#1-to-16-threads
That's a good point: you certainly could. There's some fun exploration to be done with atomic operations.
The issue is that threaded execution requires cross-origin isolation, which isn't trivial to integrate. (Example server that will serve the required headers: https://github.com/bwasti/wasmblr/blob/main/thread_example/s...)
Related posts
-
Xnnpack: High-efficiency floating-point neural network inference operators
-
Can a NPU be used for vectors?
-
[Discussion] Is XNNPACK a part of mediapipe? or should be additionally configured with mediapipe?
-
Where are Nvidia's DLSS models stored and how big are they?
-
Performance critical ML: How viable is Rust as an alternative to C++