Performance

This article sums up what to expect in terms of performance and how to optimize serverless PHP applications. The benchmarks included in this page can be reproduced via the code on GitHub.

CPU power and memory size

The main factor affecting performance is memory size. Indeed, the amount of memory is proportional to the CPU power available.

In other words, more memory means a more powerful CPU. A 1024M lambda has a CPU two times more powerful than a 512M lambda.

From 64M to 1,792M, applications run with up to one CPU (1,792M gives 1 full CPU). From 1,856M to 3,008M, applications run with 2 CPU (3,008M gives 2 full CPU). Since PHP is single-threaded and one lambda handles only 1 request at a time, using 2 CPU usually does not provide any benefit.

It is recommended to use 1024M for PHP applications, or at least to start with that. This is what Serverless deploys by default, so there is nothing to do.

To customize the amount of memory, set the memorySize option in serverless.yml:

functions:
    foo:
        handler: index.php
        # ...
        memorySize: 512 # set to 512M instead of 1024M (the default)

In the benchmark below, we run PHP's official bench.php script. This script is CPU-intensive.

128M 512M 1024M 2048M
Execution time 5.7s 1.4s 0.65s 0.33s

For comparison, bench.php runs in 1.3s on a 512M Digital Ocean server, in 0.8s on a 2.8Ghz i7 and in 0.6s on a 3.2Ghz i5. It is safe to say that a 1024M lambda provides a powerful CPU.

Costs

AWS Lambda bills the number of events + the execution time. The more memory configured for a lambda, the more expensive is the execution time.

It might be tempting to lower the memory to save money. However, a function might run slower on a smaller lambda, canceling the cost savings. For example, both of these scenarios cost the same thing:

  • a function running in 400ms on a 512M lambda
  • the same function running in 200ms (because of the faster CPU) on a 1024M lambda

In general, use smaller and slower lambdas only when speed is not important at all.

PHP runtime overhead

Bref for web apps

The FPM runtime for web apps does not add overhead to response times.

Here are execution times for an empty PHP application:

128M 512M 1024M 2048M
Execution time 10ms 1ms 1ms 1ms

Unless we use a particularly slow lambda (see the previous section, 128M is not recommended), 1ms is the same execution time when PHP runs with Apache or Nginx on a classic server.

We can see the same result with a "Hello world" written in Symfony (4ms being the minimum execution time of the framework):

128M 512M 1024M 2048M
Execution time 58ms 4ms 4ms 4ms

Bref for event-driven functions

The runtime for event-driven functions adds a small overhead:

128M 512M 1024M 2048M
Execution time 175ms 35ms 16ms 13ms

Since this runtime is often used in asynchronous scenarios (for example, processing queue messages), it is often negligible.

This overhead is caused by the PHP executable starting for every new invocation. We can skip that overhead by keeping the PHP process alive:

functions:
    hello:
        # ...
        environment:
            BREF_LOOP_MAX: 100

In the example above, the PHP process will restart only every 100 invocations, removing the overhead the rest of the time.

In that case, be careful with clearing in-memory data between every event.

Note: the PHP process will be restarted in case of a failed invocation (PHP exception thrown in the handler).

Cold starts

Code on AWS Lambda runs on-demand. When a new Lambda instance boots to handle a request, the initialization time is what we call a cold start. To learn more, you can read this article.

Bref's PHP runtimes have a cold start of 250ms on average.

This is on-par with cold starts in other language, like JavaScript, Python or Go. AWS is regularly reducing the duration of cold starts, and we are also optimizing Bref's runtimes as much as possible.

On a website with low to medium traffic, you can expect cold starts to happen for about 0.5% of the requests.

Optimizing cold starts

On small websites, cold starts can be avoided by pinging the application regularly. This keeps the lambda instances warm. Pingdom or similar services can be used, but you can also an automatic ping via serverless.yml.

While the memory size has no impact, the codebase size can increase the cold start duration. When deploying, remember to exclude assets, images, tests and any extra file in serverless.yml:

package:
    patterns:
        - '!assets/**'
        - '!node_modules/**'
        - '!tests/**'
        - ...

Read more about this in the serverless.yml documentation.

A note on VPC cold starts

Running a function inside a VPC used to induce a cold start of several seconds. This is no longer the case since October 2019.