spawner
SSVM
spawner | SSVM | |
---|---|---|
6 | 50 | |
451 | 7,952 | |
- | 1.7% | |
9.0 | 9.7 | |
over 1 year ago | 4 days ago | |
Rust | C++ | |
MIT License | Apache License 2.0 |
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
spawner
-
Container + SSH = a good development environment
For the “jhub but for any container that speaks HTTP” use case, you might find our Spawner project interesting: https://github.com/drifting-in-space/spawner
We don’t have a good story for volumes yet, but I’m open to ideas.
-
Are V8 isolates the future of computing?
Is the appeal of isolates in this case the cold start time or the isolation? We're working on some open source infrastructure for running sandboxed (gVisor) containers on the fly from web services[1], and one of the use cases people have is serving Jupyter notebooks which seems like it might resemble your use case?
[1] https://github.com/drifting-in-space/spawner/
-
Fly Machines: An API for Fast-Booting VMs
yes! a fellow HN user e-mailed me about his project "Spawner"
https://github.com/drifting-in-space/spawner
check out the demo: https://www.youtube.com/watch?v=aGsxxcQRKa4
-
Ask HN: Who is hiring? (May 2022)
Drifting in Space | Full-time | NYC | https://driftingin.space
We make Jamsocket (https://jamsocket.com/), which allows application developers to spin up and connect to server-side compute. This allows browser-based applications to do computationally-intensive things that are otherwise impossible in the browser.
We went through YC and just raised a seed round and are looking to build up our team. We are based in NYC but are open to remote for experience candidates.
Our tech stack includes Rust, NATS, Docker, Postgres, TypeScript. We have lots of fun technical problems that get into the nitty-gritty of networking and operating systems. We are excited to build a diverse team and encourage non-traditional candidates to apply.
Email [email protected] or see more details here: https://www.ycombinator.com/companies/drifting-in-space/jobs...
-
Launch HN: Drifting in Space (YC W22) – A server process for every user
Hi HN, we’re Paul and Taylor, and we’re launching Drifting in Space (https://driftingin.space). We build server software for performance-intensive browser-based applications. We make it easy to give every user of your app a dedicated server-side process, which starts when they open your application and stops when they close the tab.
Many high-end web apps give every user a dedicated connection to a server-side process. That is how they get the low latency that you need for ambitious products like full-fledged video editing tools and IDEs. This is hard for smaller teams to recreate, though, because it takes a significant ongoing engineering investment. That’s where we come in—we make this architecture available to everyone, so you can focus on your app instead of its infrastructure. You can think of it like Heroku, except that each of your users gets their own server instance.
I realized that something like this was needed while working on data-intensive tools at a hedge fund. I noticed that almost all new application software, whether it was built in-house or third-party SaaS, was delivered as a browser application rather than native. Although browsers are more powerful than ever, I knew from experience that industrial-scale data-heavy apps posed problems, because neither the browser or a traditional stateless server architecture could provide the compute resources needed for low-latency interaction with large datasets. I began talking about this with my friend Taylor, who had encountered similar limitations while working on data analysis and visualization tools at Datadog and Uber. We decided to team up and build a company around solving it.
We have two products, an open source package and a managed platform. Spawner, the open source part, provides an API for web apps to spawn a session-lived process. It manages the process’s lifecycle, exposing it over HTTPS, tracking inbound connections, and shutting it down when it becomes idle (i.e. when the user closes their tab). It’s open source (MIT) and available at https://github.com/drifting-in-space/spawner.
Jamsocket is our managed platform, which uses Spawner internally. It provides the same API, but frees you from having to deal with any cluster or network configuration to ship code. From an app developer’s point of view, using it is similar to using platforms like Netlify or Render. You stay in the web stack and never have to touch Kubernetes.
Here's an example. Imagine you make an application for investigating fraud in a large transaction database. Users want to interactively filter, aggregate, and visualize gigabytes of transactions as a graph. Instead of sending all of the data down to the browser and doing the work there, you would put your code in a container and upload it to our platform. Then, whenever a fraud analyst opens your application, you hit an API we provide to spin up a dedicated backend for that analyst. Your browser code then opens a WebSocket connection directly to that backend, which it uses to stream data as the analyst applies filters or zooms/pans the visualization.
We're different from most managed platforms because we give each user a dedicated process. That said, there are a few other services that do run long-lived processes for each user. Architecturally, we're most similar to Agones. Agones is targeted at games where the client can speak UDP to an arbitrary IP; we target applications that want to connect directly from browsers to a hostname over HTTPS. In the Erlang world, the OTP stack provides similar functionality, but you have to embrace Erlang/Elixir to get the benefits of it; we are entirely language-agnostic. Cloudflare Durable Objects support a form of long-lived processes, but are focused on use cases around program state synchronization rather than arbitrary high-compute/memory use cases.
We have a usage-based billing model, similar to Heroku. We charge you for the compute you use and take a cut. Usage billing scales to zero, so it’s approachable for weekend experiments. We have not solidified a price plan yet, but we’re aiming to provide an instance capable of running VS Code (as an example) for about 10 cents an hour, fractionally metered. High-memory and high-CPU backends will cost more, and heavy users will get volume discounts. Our target customers are desktop-like SaaS apps and internal data tools.
As mentioned, our core API is open source and available at https://github.com/drifting-in-space/spawner. The managed platform is in beta and we’re currently onboarding users from a waitlist, to make sure that we have the server capacity to scale. If you’re interested, you’re welcome to sign up for it here: https://driftingin.space.
Have you built a similar infrastructure for your application? We’re always interested in hearing the approaches people have already taken to this problem and learning what their pain points are.
SSVM
- A WASM runtime for running LLMs locally
-
Time-series data ingestion from Rust WebAssembly application, leveraging GreptimeDB and WasmEdge
WasmEdge GitHub address: https://github.com/WasmEdge/WasmEdge.
-
Orca-2-13B Runs Directly on Rust+WASM – No Python/C++ Hassles
I see that they recently changed the intro of WasmEdge on Github [1] to " WasmEdge is the easiest and fastest way to run LLMs on your own devices. "
Since it's a wasm runtime capable of many things I find bizarre that they now start describing it with a ultra-specific use case
- [1] https://github.com/WasmEdge/WasmEdge
-
Run LLMs on my own Mac fast and efficient Only 2 MBs
Mmm…
The wasm-nn that this relies on (https://github.com/WebAssembly/wasi-nn) is a proposal that relies of arbitrary plugin backends sending arbitrarily chunks to some vendor implementation. The api is literally like set input, compute, set output.
…and that is totally non portable.
The reason this works, is because it’s relying on the abstraction already implemented in llama.cpp that allows it to take a gguf model and map it to multiple hardware targets,which you can see has been lifted here: https://github.com/WasmEdge/WasmEdge/tree/master/plugins/was...
So..
> Developers can refer to this project to write their machine learning application in a high-level language using the bindings, compile it to WebAssembly, and run it with a WebAssembly runtime that supports the wasi-nn proposal, such as WasmEdge.
Is total rubbish; no, you can’t.
This isn’t portable.
It’s not sandboxed.
If you have a wasm binary you might be able to run it if the version of the runtime you’re using happens to implement the specific ggml backend you need, which it probably doesn’t… because there’s literally no requirement for it to do so.
There’s a lot of “so portable” talk in this article which really seems misplaced.
- Security Slam 2023: Contribute to WasmEdge and Elevate Open Source Security
-
Requiem for a Stringref
WasmEdge isn't there yet: https://github.com/WasmEdge/WasmEdge/issues/1122#issuecommen...
- Should You Be Scared of Unix Signals?
- WasmEdge 0.13.0: Unified CLI, ARM Support and Migrating Extensions to Plugins
-
ChatGPT-powered code review bot to boost your PR merge. Deploy in 5 mins
Example 1: Analyze the content and risks of each commit in the PR. Then make a summary. https://github.com/WasmEdge/WasmEdge/pull/2394#issuecomment-...
-
Cloud, Why So Difficult?
There has also been a few "cloud-native" runtimes based on WASM, like WasmEdge but there's a few others (can't remember their names!)...
What are some alternatives?
splitter - React component for building split views like in VS Code
wasmer - 🚀 The leading Wasm Runtime supporting WASIX, WASI and Emscripten
stateroom - A lightweight framework for building WebSocket-based application backends.
wasmtime - A fast and secure runtime for WebAssembly
blueboat - All-in-one, multi-tenant serverless JavaScript runtime.
aws-lambda-wasm-runtime - A template project for building high-performance, portable, and safe serverless functions in AWS Lambda.
wizer - The WebAssembly Pre-Initializer
WAVM - WebAssembly Virtual Machine
ContainerSSH - ContainerSSH: Launch containers on demand
dapr-wasm - A template project to demonstrate how to run WebAssembly functions as sidecar microservices in dapr
llvm-project - The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.
WasmEdge-go - The GO language SDK and API for WasmEdge