llrt
winterjs
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.
llrt
-
Show HN: Self-Host Next.js in Production
Any plans to add support for? https://github.com/awslabs/llrt
It would also be nice to have a V8/deno/bun based edge hosting option that supports the Next.js edge and middleware code splitting. That's the missing piece for most homebrew "edge" setups. Production CDNs like Clouflare and Supabase all offer this.
-
Everything Suffers from Cold Starts
Vlad Ionescu: Scaling containers on AWS in 2022 GitHub: awslabs/llrt AWS Documentation: Understanding the Lambda execution environment Amazon Science: How AWS's Firecracker virtual machines work Lumigo GitHub: MiddyJS
-
Porffor: A from-scratch experimental ahead-of-time JS engine
Its refreshing to see all the various JS engines that are out there for various usecases.
I have been working on providing quickjs with more node compatible API through llrt [1] for embedding into applications for plugins.
[1] https://github.com/awslabs/llrt
-
[Lab] AWS Lambda LLRT vs Node.js
AWS has open-sourced its JavaScript runtime, called LLRT (Low Latency Runtime), an experimental, lightweight JavaScript runtime designed to address the growing demand for fast and efficient Serverless applications.
-
Unlocking Next-Gen Serverless Performance: A Deep Dive into AWS LLRT
FROM --platform=arm64 busybox WORKDIR /var/task/ COPY app.mjs ./ ADD https://github.com/awslabs/llrt/releases/latest/download/llrt-container-arm64 /usr/bin/llrt RUN chmod +x /usr/bin/llrt ENV LAMBDA_HANDLER "app.handler" CMD [ "llrt" ]
-
Is AWS Lambda Cold Start Still an Issue?
Let’s get the simplest use case out of the way: cases where the cold starts are so fast that it’s not an issue for you. That’s usually the case for function that use runtimes such as C++, Go, Rust, and LLRT. However, you must follow the best practices and optimizations in every runtime to maintain a low impact cold start.
-
JavaScript News, Updates, and Tutorials: February 2024 Edition
But compared to other runtimes, LLRT is not so good in terms of performance when it comes to dealing with large data processing, Monte Carlo simulations, or performing tasks with a large number of iterations. The AWS team says that it is best suited for working with smaller Serverless functions dedicated to tasks such as data transformation, real-time processing, AWS service integrations, authorization, validation, etc. Visit the GitHub repository of this project to learn more information.
- FLaNK Stack 26 February 2024
-
People Matter more than Technology when Building Serverless Applications
And lastly, lean into your cloud vendor. Stop trying to build a better mouse trap. Advances in technology are happening all the time. The speed of AWS' Lambda has been rapidly improving over the past couple of years with the launch of things like SnapStart and LLRT
- Hono v4.0.0
winterjs
-
LLRT: A low-latency JavaScript runtime from AWS
LLRT is built using Rust, Tokio and QuickJS under the hood.
We (at Wasmer), have been working on a very similar approach: WinterJS, which is using SpiderMonkey instead (but also built on top of Rust and Tokio). https://github.com/wasmer/winterjs
What are some alternatives?
winterjs - Winter is coming... ❄️
pljs - PLJS - Javascript Language Plugin for PostreSQL
mud-pi - A simple MUD server in Python, for teaching purposes, which could be run on a Raspberry Pi
workerd - The JavaScript / Wasm runtime that powers Cloudflare Workers
hermes - A JavaScript engine optimized for running React Native.
sst - Build full-stack apps on your own infrastructure.
hono - Web framework built on Web Standards
h3 - ⚡️ Minimal H(TTP) framework built for high performance and portability