Home > Web Front-end > JS Tutorial > Encore.ts — Faster cold starts than NestJS & Fastify

Encore.ts — Faster cold starts than NestJS & Fastify

WBOY
Release: 2024-09-05 06:38:03
Original
673 people have browsed it

A couple of months ago we released Encore.ts — an Open Source backend framework for TypeScript.

Since there are already many frameworks out there, we want to share some of the outlier design decisions we've made and how they lead to remarkable performance numbers.

We recently published performance benchmarks showing that Encore.ts achieves 9x request throughput compared to Express.js, and 2x compared to Fastify.

Today, we're continuing on our performance journey by diving into how Encore.ts achieves incredibly fast cold start startup times.

Performance benchmarks

This time we've benchmarked Encore.ts, Fastify, NestJS and Express to see how each framework performs when it comes to cold startup times.

The benchmark program registers 10 API endpoints, each with a simple schema, and sets up schema validation.
For schema validation we used Zod where possible.
In the case of Fastify we used Ajv as the officially supported schema validation library.

We measured the time from when JavaScript code begins executing until the server is ready to accept incoming requests.
For each benchmark we took the best result of five runs.

Enough talk, let's dig into the numbers!

Encore.ts cold starts are 17x faster than NestJS and Fastify

Encore.ts —  Faster cold starts than NestJS & Fastify

(Check out the benchmark code on GitHub.)

As you can see, Encore.ts achieves remarkable fast cold startup times, over 5x faster than Express and over 17x faster than NestJS.

How is this possible? From our testing we've identified two major sources of performance, both related to how Encore.ts works under the hood.

But before we get there, let's talk about what cold starts really are, and why they matter.

What is a cold start?

In the context of serverless, a cold start is when the underlying platform first needs to spins up a new instance of your server in order to serve an incoming request. (It can also refer to the first time a new instance of your server is started up to handle a request, for example after a deployment.)

Since the request is effectively on hold until the process starts up and is ready to handle the request, reducing cold startup times can have a large impact on the long-tail latency of your application.

This is especially important for distributed systems where you have multiple serverless functions, as it's much more likely you will encounter a cold start in some part of the system when handling a request.

The anatomy of a cold start

Exactly what happens during a cold start depends a bit on the platform you're deploying to (Kubernetes, Lambda, Cloud Run, etc.).
But in general, the process looks something like this:

  1. Platform downloads the code/container image for the serverless function
  2. Platform spins up a new instance of the container/serverless function/container
  3. The container/function initializes itself (importing JavaScript modules, running initialization code, etc.)

After these initialization steps the cold start is complete, and the serverless function begins processing the incoming request.

The first two steps are largely out of our control (other than by making sure the size of the code/container is optimized), so let's focus our attention on the third step.

In fact, let's further break down the third step, and assuming we're running Node.js:

  1. The node process starts up and begins initializing the V8 JavaScript engine
  2. The entrypoint file is parsed, loaded, and begins executing application code
  3. When the JavaScript code executes import and require statements, yet more files are loaded, parsed and executed. (Repeat many times for applications with lots of dependencies.)

Finally, after all dependencies have been loaded and all the initialization code has executed, the container/serverless function is ready to handle incoming requests.

Optimizing cold starts

The breakdown above gives us clear targets for optimization, and Encore.ts heavily optimizes all the steps it has control over.

Optimization 1: Rust runtime

Encore.ts is implemented in Rust and loaded into Node.JS as a native module. This has several benefits for cold starts:

Less JavaScript to parse and execute. Since JavaScript is an interpreted language, all JavaScript code needs to be read from disk, parsed, and executed. Encore.ts, as a pre-compiled native module, loads extremely quickly and doesn't need to be parsed or executed by the JavaScript engine (V8).

Zero NPM dependencies. Since Encore.ts implements all its functionality using Rust, it has no NPM dependencies whatsoever, which further reduces the amount of JavaScript that needs to be executed during a cold start.

Pre-compiled and optimized. JavaScript relies heavily on just-in-time compilation (JIT), where code that gets executed repeatedly gets optimized by the JavaScript engine. This makes a lot of sense for an interpreted language, but it also means that execution is quite a bit slower the first time a piece of code runs, which impacts cold starts considerably. Since Encore.ts is implemented in Rust, it's pre-compiled and heavily optimized for the platform it's running on, which means it's fast from the first time it's executed.

Optimization 2: Efficient Docker images

Encore.ts by default builds minified Docker images, by only including transpiled JavaScript and the necessary dependencies to run the application. This reduces bundle sizes, which in turn reduces the time it takes to download and start up the container.

Additionally, several compute platforms have added support for streaming Docker images, which means that the platform can start the container before the entire image has been downloaded. Encore.ts has built-in support for this, and automatically prioritizes the parts of the image that are needed to reduce cold starts.

Wrapping up

By combining a Rust runtime with optimized Docker images, Encore.ts is able to achieve remarkable cold start times, which can have a large impact on the long-tail latency of your application.

If performance matters to your project, it might be a good idea to try out Encore.ts.

And it's all Open Source, so you can check out the code and contribute on GitHub.

Or just give it a try and let us know what you think!

The above is the detailed content of Encore.ts — Faster cold starts than NestJS & Fastify. For more information, please follow other related articles on the PHP Chinese website!

source:dev.to
Statement of this Website
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
Popular Tutorials
More>
Latest Downloads
More>
Web Effects
Website Source Code
Website Materials
Front End Template