wizer
SSVM
Our great sponsors
wizer | SSVM | |
---|---|---|
10 | 50 | |
875 | 7,932 | |
3.4% | 3.3% | |
7.0 | 9.8 | |
11 days ago | 4 days ago | |
Rust | C++ | |
Apache License 2.0 | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
wizer
-
RustPython
> once by the wasm runtime to compile the rust-python wasm
I'm not sure what you mean by that. The runtime doesn't compile WASM, it simply executes it.
There are tools for dealing with interpreter runtime overhead this by pre-initalizing the environment like Wizer[0]. ComponentizeJS[1] uses it to pre-initialize the Spidermoney engine it packages to gain fast startup times (and you can then prune the initialization only code with wasm-opt). As techniques like ComponentizeJS are also being applied for a specific set of interpreted files, you can even prune parts of the interpreter that would never be used for that specific program. If you want to go even further you could record specific execution profiles and optimize further by those.
[0]: https://github.com/bytecodealliance/wizer
[1]: https://github.com/bytecodealliance/ComponentizeJS
- Are V8 isolates the future of computing?
-
Netlify Edge Functions: A new serverless runtime powered by Deno
Edge functions are typically run intermittently, with their runtime stopped to free up resources between runs. Therefore a big factor is startup and shutdown speed. Containers are pretty bad there. Deno is better, and WASM is unbeatable, especially with things like Wizer[0].
[0]https://github.com/bytecodealliance/wizer
-
Building a WebAssembly-powered serverless platform
I imagine startup cost could be amortized by something like wizer: https://github.com/bytecodealliance/wizer
-
Containerless! How to Run WebAssembly Workloads on Kubernetes with Rust
There are security benefits to running each request in its own instance, as it helps prevent accidental leaking of state between requests. To avoid doing lots of expensive initializations, we have a tool called wizer which lets users run their program's initialization once, create a snapshot, and then use that snapshot to do fast startups that don't rerun the whole initialization each time.
-
Is it possible in Rust to save the complete state of a program and restore it later? Such as may be accomplished in some implementations of Common Lisp
See https://github.com/bytecodealliance/wizer for an implementation of this approach.
-
Bytecode Alliance
It should probably be named "Making JavaScript to startup fast on WebAssembly", since the runtime speed is not really improved by the approach they exposed.
Besides that I think Wizer [1] is both an elegant and a simple solution to speed up startup speed with Wasm.
[1] - https://github.com/bytecodealliance/wizer#using-wizer-as-a-l...
-
A JavaScript optimizing compiler
A similar project, for WebAssembly so with limited scope is this: https://github.com/bytecodealliance/wizer. And somehow similar but limited on LLVM IR a colleague worked on this for Cheerp (the compiler used here as backend): https://github.com/leaningtech/cheerp-meta/wiki/Cheerp-PreExecuter.
- Wizer: snapshot an initialized Wasm instance and save the result as a new, pre-initialized Wasm module. Up to 6x faster start up on my test workloads
- Wiser: snapshot an initialized Wasm instance and save the result as a new, pre-initialized Wasm module. Up to 6x faster start up on my test workloads
SSVM
- A WASM runtime for running LLMs locally
-
Time-series data ingestion from Rust WebAssembly application, leveraging GreptimeDB and WasmEdge
WasmEdge GitHub address: https://github.com/WasmEdge/WasmEdge.
-
Orca-2-13B Runs Directly on Rust+WASM – No Python/C++ Hassles
I see that they recently changed the intro of WasmEdge on Github [1] to " WasmEdge is the easiest and fastest way to run LLMs on your own devices. "
Since it's a wasm runtime capable of many things I find bizarre that they now start describing it with a ultra-specific use case
- [1] https://github.com/WasmEdge/WasmEdge
-
Run LLMs on my own Mac fast and efficient Only 2 MBs
Mmm…
The wasm-nn that this relies on (https://github.com/WebAssembly/wasi-nn) is a proposal that relies of arbitrary plugin backends sending arbitrarily chunks to some vendor implementation. The api is literally like set input, compute, set output.
…and that is totally non portable.
The reason this works, is because it’s relying on the abstraction already implemented in llama.cpp that allows it to take a gguf model and map it to multiple hardware targets,which you can see has been lifted here: https://github.com/WasmEdge/WasmEdge/tree/master/plugins/was...
So..
> Developers can refer to this project to write their machine learning application in a high-level language using the bindings, compile it to WebAssembly, and run it with a WebAssembly runtime that supports the wasi-nn proposal, such as WasmEdge.
Is total rubbish; no, you can’t.
This isn’t portable.
It’s not sandboxed.
If you have a wasm binary you might be able to run it if the version of the runtime you’re using happens to implement the specific ggml backend you need, which it probably doesn’t… because there’s literally no requirement for it to do so.
There’s a lot of “so portable” talk in this article which really seems misplaced.
- Security Slam 2023: Contribute to WasmEdge and Elevate Open Source Security
-
Requiem for a Stringref
WasmEdge isn't there yet: https://github.com/WasmEdge/WasmEdge/issues/1122#issuecommen...
- Should You Be Scared of Unix Signals?
- WasmEdge 0.13.0: Unified CLI, ARM Support and Migrating Extensions to Plugins
-
ChatGPT-powered code review bot to boost your PR merge. Deploy in 5 mins
Example 1: Analyze the content and risks of each commit in the PR. Then make a summary. https://github.com/WasmEdge/WasmEdge/pull/2394#issuecomment-...
-
Cloud, Why So Difficult?
There has also been a few "cloud-native" runtimes based on WASM, like WasmEdge but there's a few others (can't remember their names!)...
What are some alternatives?
quickjs-emscripten - Safely execute untrusted Javascript in your Javascript, and execute synchronous code that uses async functions
wasmer - 🚀 The leading Wasm Runtime supporting WASIX, WASI and Emscripten
TablaM - The practical relational programing language for data-oriented applications
wasmtime - A fast and secure runtime for WebAssembly
aws-lambda-wasm-runtime - A template project for building high-performance, portable, and safe serverless functions in AWS Lambda.
go - The Go programming language
WAVM - WebAssembly Virtual Machine
go-wasm-bake - Experimenting with eager evaluation of Go WASM code
dapr-wasm - A template project to demonstrate how to run WebAssembly functions as sidecar microservices in dapr
wagi - Write HTTP handlers in WebAssembly with a minimal amount of work
WasmEdge-go - The GO language SDK and API for WasmEdge