Examples
A collection of example apps for reference.
Below is a collection of example SST apps. These are available in the examples/
directory of the repo.
The descriptions for these examples are generated using the comments in the sst.config.ts
of the app.
Contributing
To contribute an example or to edit one, submit a PR to the repo.
Make sure to document the sst.config.ts
in your example.
API Gateway auth
Enable IAM and JWT authorizers for API Gateway routes.
View the full example.
AWS Astro container with Redis
Creates a hit counter app with Astro and Redis.
This deploys Astro as a Fargate service to ECS and it’s linked to Redis.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:4321
you’ll see a counter update as you refresh the page.
Finally, you can deploy it by adding the Dockerfile
that’s included in this example and
running npx sst deploy --stage production
.
View the full example.
AWS Astro streaming
Follows the Astro Streaming guide to create an app that streams HTML.
The responseMode
in the astro-sst
adapter
is set to enable streaming.
Now any components that return promises will be streamed.
You should see the friends section load after a 3 second delay.
Safari uses a different heuristic to determine when to stream data. You need to render enough initial HTML to trigger streaming. This is typically only a problem for demo apps.
There’s nothing to configure for streaming in the Astro
component.
View the full example.
Bucket policy
Create an S3 bucket and transform its bucket policy.
View the full example.
Bucket queue notifications
Create an S3 bucket and subscribe to its events with an SQS queue.
View the full example.
Bucket notifications
Create an S3 bucket and subscribe to its events with a function.
View the full example.
Bucket topic notifications
Create an S3 bucket and subscribe to its events with an SNS topic.
View the full example.
AWS Bun Elysia container
Deploys a Bun Elysia API to AWS.
You can get started by running.
Now you can add a service.
Start your app locally.
This example lets you upload a file to S3 and then download it.
Finally, you can deploy it using bun sst deploy --stage production
.
View the full example.
AWS Bun Redis
Creates a hit counter app with Bun and Redis.
This deploys Bun as a Fargate service to ECS and it’s linked to Redis.
We also have a couple of scripts. A dev
script with a watcher and a build
script
that used when we deploy to production.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:3000
you’ll see a counter update as you refresh the page.
Finally, you can deploy it using bun sst deploy --stage production
using a Dockerfile
that’s included in the example.
View the full example.
AWS Cluster private service
Adds a private load balancer to a service by setting the loadBalancer.public
prop to
false
.
This allows you to create internal services that can only be accessed inside a VPC.
View the full example.
AWS Cluster with API Gateway
Expose a service through API Gateway HTTP API using a VPC link.
This is an alternative to using a load balancer. Since API Gateway is pay per request, it works out a lot cheaper for services that don’t get a lot of traffic.
You need to specify which port in your service will be exposed through API Gateway.
Your API Gateway HTTP API also needs to be in the same VPC as the service.
View the full example.
Subscribe to queues
Create an SQS queue, subscribe to it, and publish to it from a function.
View the full example.
AWS Deno Redis
Creates a hit counter app with Deno and Redis.
This deploys Deno as a Fargate service to ECS and it’s linked to Redis.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:8000
you’ll see a counter update as you refresh the page.
Finally, you can deploy it using sst deploy --stage production
using a Dockerfile
that’s included in the example.
View the full example.
DynamoDB streams
Create a DynamoDB table, enable streams, and subscribe to it with a function.
View the full example.
EC2 with Pulumi
Use raw Pulumi resources to create an EC2 instance.
View the full example.
AWS EFS with SQLite
Mount an EFS file system to a function and write to a SQLite database.
The file system is mounted to /mnt/efs
in the function.
This example is for demonstration purposes only. It’s not recommended to use EFS for databases in production.
View the full example.
AWS EFS with SurrealDB
We use the SurrealDB docker image to run a server in a container and use EFS as the file system.
We then connect to the server from a Lambda function.
This uses the SurrealDB client to connect to the server.
This example is for demonstration purposes only. It’s not recommended to use EFS for databases in production.
View the full example.
AWS EFS
Mount an EFS file system to a function and a container.
This allows both your function and the container to access the same file system. Here they both update a counter that’s stored in the file system.
The file system is mounted to /mnt/efs
in both the function and the container.
View the full example.
AWS Express Redis
Creates a hit counter app with Express and Redis.
This deploys Express as a Fargate service to ECS and it’s linked to Redis.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:80
you’ll see a counter update as you refresh the page.
Finally, you can deploy it using npx sst deploy --stage production
using a Dockerfile
that’s included in the example.
View the full example.
FFmpeg in Lambda
Uses FFmpeg to process videos. In this example, it takes a clip.mp4
and grabs a single frame from it.
We use the ffmpeg-static
package that
contains pre-built binaries for all architectures.
We can use this to spawn a child process and run FFmpeg.
We don’t need a layer when we deploy this because SST will use the right binary for the
target Lambda architecture; including arm64
.
All this is handled by nodejs.install
.
View the full example.
AWS Hono container with Redis
Creates a hit counter app with Hono and Redis.
This deploys Hono API as a Fargate service to ECS and it’s linked to Redis.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:3000
you’ll see a counter update as you refresh the page.
Finally, you can deploy it by:
-
Using the
Dockerfile
that’s included in this example. -
This compiles our TypeScript file, so we’ll need add the following to the
tsconfig.json
. -
Install TypeScript.
-
And add a
build
script to ourpackage.json
.
And finally, running npx sst deploy --stage production
.
View the full example.
AWS Hono streaming
An example on how to enable streaming for Lambda functions using Hono.
While sst dev
doesn’t support streaming, we can conditionally enable it on deploy.
This will return the standard handler for sst dev
.
To test this in your terminal, use the curl
command with the --no-buffer
option.
Here we are using a Function URL directly because API Gateway doesn’t support streaming.
View the full example.
IAM permissions boundaries
Use permissions boundaries to set the maximum permissions for all IAM roles that’ll be created in your app.
In this example, the Function has the s3:ListAllMyBuckets
and sqs:ListQueues
permissions. However, we create a permissions boundary that only allows s3:ListAllMyBuckets
.
And we apply it to all Roles in the app using the global
$transform
.
As a result, the Function is only allowed to list S3 buckets. If you open the deployed URL, you’ll see that the SQS list call fails.
Learn more about AWS IAM permissions boundaries.
View the full example.
Current AWS account
You can use the aws.getXXXXOutput()
provider functions to get info about the current
AWS account.
Learn more about provider functions.
View the full example.
AWS JSX Email
Uses JSX Email and the Email
component to design and send emails.
To test this example, change the sst.config.ts
to use your own email address.
Then run.
You’ll get an email from AWS asking you to confirm your email address. Click the link to verify it.
Next, go to the URL in the sst dev
CLI output. You should now receive an email rendered
using JSX Email.
Once you are ready to go to production, you can:
- Request production access for SES
- And use your domain to send emails
View the full example.
Kinesis streams
Create a Kinesis stream, and subscribe to it with a function.
View the full example.
AWS Lambda Go
This example shows how to use the go
runtime in your Lambda
functions.
Our Go function is in the src
directory and we point to it in our function.
We are also linking it to an S3 bucket. We can reference the bucket in our function.
The resource.Get
function is from the SST Go SDK.
The sst dev
CLI also supports running your Go function Live.
View the full example.
AWS Lambda retry with queues
An example on how to retry Lambda invocations using SQS queues.
Create a SQS retry queue which will be set as the destination for the Lambda function.
Create a bus subscriber which will publish messages to the bus. Include a DLQ for messages that continue to fail.
The Retry function will read mesaages and send back to the queue to be retried with a backoff.
View the full example.
AWS Lambda streaming
An example on how to enable streaming for Lambda functions.
While sst dev
doesn’t support streaming, you can use the
lambda-stream
package to test locally.
Then, you can use the streamifyResponse
function to wrap your handler:
When deployed, this will use the awslambda.streamifyResponse
.
To test this in your terminal, use the curl
command with the --no-buffer
option.
Here we are using a Function URL directly because API Gateway doesn’t support streaming.
View the full example.
AWS Lambda in a VPC
You can use SST to locally work on Lambda functions that are in a VPC. To do so, you’ll
need to enable bastion
and nat
on the Vpc
component.
The NAT gateway is necessary to allow your Lambda function to connect to the internet. While, the bastion host is necessary for your local machine to be able to tunnel to the VPC.
You’ll need to install the tunnel, if you haven’t done this before.
This needs sudo to create the network interface on your machine. You’ll only need to do this once.
Now you can run sst dev
, your function can access resources in the VPC. For example, here
we are connecting to a Redis cluster.
The Redis cluster is in the same VPC as the function.
View the full example.
AWS multi-region
To deploy resources to multiple AWS regions, you can create a new provider for the region you want to deploy to.
And then pass that in to the resource.
If no provider is passed in, the default provider will be used. And if no region is specified, the default region from your credentials will be used.
View the full example.
AWS Next.js add behavior
Here’s how to add additional routes or cache behaviors to the CDN of a Next.js app deployed with OpenNext to AWS.
Specify the path pattern that you want to forward to your new origin. For example, to forward
all requests to the /blog
path to a different origin.
And then specify the domain of the new origin.
We use this to transform
our site’s CDN and add the additional behaviors.
View the full example.
AWS Next.js basic auth
Deploys a simple Next.js app and adds basic auth to it.
This is useful for dev environments where you want to share your app your team but ensure that it’s not publicly accessible.
This works by injecting some code into a CloudFront function that checks the basic auth
header and matches it against the USERNAME
and PASSWORD
secrets.
To deploy this, you need to first set the USERNAME
and PASSWORD
secrets.
If you are deploying this to preview environments, you might want to set the secrets using
the --fallback
flag.
View the full example.
AWS Next.js container with Redis
Creates a hit counter app with Next.js and Redis.
This deploys Next.js as a Fargate service to ECS and it’s linked to Redis.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:3000
you’ll see a counter update as you refresh the page.
Finally, you can deploy it by:
- Setting
output: "standalone"
in yournext.config.mjs
file. - Adding a
Dockerfile
that’s included in this example. - Running
npx sst deploy --stage production
.
View the full example.
AWS Next.js streaming
An example of how to use streaming Next.js RSC. Uses Suspense
to stream an async component.
For this demo we also need to make sure the route is not statically built.
This is deployed with OpenNext, which needs a config to enable streaming.
You should see the friends section load after a 3 second delay.
Safari uses a different heuristic to determine when to stream data. You need to render enough initial HTML to trigger streaming. This is typically only a problem for demo apps.
View the full example.
AWS Postgres local
In this example, we connect to a locally running Postgres instance for dev. While on deploy, we use RDS.
We use the docker run
CLI
to start a local container with Postgres. You don’t have to use Docker, you can use
Postgres.app or any other way to run Postgres locally.
The data is saved to the .sst/storage
directory. So if you restart the dev server, the
data will still be there.
We then configure the dev
property of the Postgres component with the settings for the
local Postgres instance.
By providing the dev
prop for Postgres, SST will use the local Postgres instance and
not deploy a new RDS database when running sst dev
.
It also allows us to access the database through a Reosurce link
without having to
conditionally check if we are running locally.
The above will work in both sst dev
and sst deploy
.
View the full example.
Prisma in Lambda
To use Prisma in a Lambda function you need to
- Generate the Prisma Client with the right architecture
- Copy the generated client to the function
- Run the function inside a VPC
You can set the architecture using the binaryTargets
option in prisma/schema.prisma
.
You can also switch to ARM, just make sure to also change the function architecture in your
sst.config.ts
.
To generate the client, you need to run prisma generate
when you make changes to the
schema.
Since this needs to be done on every deploy, we add a postinstall
script to the package.json
.
This runs the command on npm install
.
We then need to copy the generated client to the function when we deploy.
Our function also needs to run inside a VPC, since Prisma doesn’t support the Data API.
Prisma in serverless environments
Prisma is not great in serverless environments. For a couple of reasons:
- It doesn’t support Data API, so you need to manage the connection pool on your own.
- Without the Data API, your functions need to run inside a VPC.
- You cannot use
sst dev
without connecting to the VPC.
- You cannot use
- Due to the internal architecture of their client, it’s also has slower cold starts.
Instead we recommend using Drizzle. This example is here for reference for people that are already using Prisma.
View the full example.
Puppeteer in Lambda
To use Puppeteer in a Lambda function you need:
puppeteer-core
- Chromium
- In
sst dev
, we’ll use a locally installed Chromium version. - In
sst deploy
, we’ll use the@sparticuz/chromium
package. It comes with a pre-built binary for Lambda.
- In
Chromium version
Since Puppeteer has a preferred version of Chromium, we’ll need to check the version of Chrome that a given version of Puppeteer supports. Head over to the Puppeteer’s Chromium Support page and check which versions work together.
For example, Puppeteer v23.1.1 supports Chrome for Testing 127.0.6533.119. So, we’ll use the
v127 of @sparticuz/chromium
.
Install Chromium locally
To use this locally, you’ll need to install Chromium.
Once installed you’ll see the location of the Chromium binary, /tmp/localChromium/chromium/mac_arm-1350406/chrome-mac/Chromium.app/Contents/MacOS/Chromium
.
Update this in your Lambda function.
You’ll notice we are using the right binary with the SST_DEV
environment variable.
Deploy
We don’t need a layer to deploy this because @sparticuz/chromium
comes with a pre-built
binary for Lambda.
We just need to set it in the nodejs.install
.
And on deploy, SST will use the right binary.
We are giving our function more memory and a longer timeout since running Puppeteer can take a while.
View the full example.
Subscribe to queues
Create an SQS queue, subscribe to it, and publish to it from a function.
View the full example.
AWS Redis local
In this example, we connect to a local Docker Redis instance for dev. While on deploy, we use Redis ElastiCache.
We use the docker run
CLI
to start a local Redis server. You don’t have to use Docker, you can run it locally any way
you want.
The data is persisted to the .sst/storage
directory. So if you restart the dev server,
the data will still be there.
We then configure the dev
property of the Redis component with the settings for the
local Redis server.
By providing the dev
prop for Redis, SST will use the local Redis server and
not deploy a new Redis ElastiCache cluster when running sst dev
.
It also allows us to access Redis through a Reosurce link
.
The local Redis server is running in standalone
mode, whereas on deploy it’ll be in
cluster
mode. So our Lambda function needs to connect using the right config.
View the full example.
AWS Remix container with Redis
Creates a hit counter app with Remix and Redis.
This deploys Remix as a Fargate service to ECS and it’s linked to Redis.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:5173
you’ll see a counter update as you refresh the page.
Finally, you can deploy it by adding the Dockerfile
that’s included in this example and
running npx sst deploy --stage production
.
View the full example.
AWS Remix streaming
Follows the Remix Streaming guide to create an app that streams data.
Uses the defer
utility to stream data through the loader
function.
Then uses the the Suspense
and Await
components to render the data.
You should see the friends section load after a 3 second delay.
Safari uses a different heuristic to determine when to stream data. You need to render enough initial HTML to trigger streaming. This is typically only a problem for demo apps.
Streaming works out of the box with the Remix
component.
View the full example.
Router and bucket
Creates a router that serves static files from the public
folder of a given bucket.
View the full example.
Router and function URL
Creates a router that routes all requests to a function with a URL.
View the full example.
AWS Cluster Service Discovery
In this example, we are connecting to a service running on a cluster using its AWS Cloud Map service host name. This is useful for service discovery.
We are deploying a service to a cluster in a VPC. And we can access it within the VPC using the service’s cloud map hostname.
Here we are accessing it through a Lambda function that’s linked to the service and is deployed to the same VPC.
View the full example.
Sharp in Lambda
Uses the Sharp library to resize images. In this example,
it resizes a logo.png
local file to 100x100 pixels.
We don’t need a layer to deploy this because sharp
comes with a pre-built binary for Lambda.
This is handled by nodejs.install
.
In dev, this uses the sharp npm package locally.
On deploy, SST will use the right binary from the sharp package for the target Lambda architecture.
View the full example.
AWS SolidStart WebSocket endpoint
Deploys a SolidStart app with a WebSocket endpoint in a container to AWS.
Uses the experimental WebSocket support in Nitro.
Once deployed you can test the /ws
endpoint and it’ll send a message back after a 3s delay.
View the full example.
AWS static site basic auth
This deploys a simple static site and adds basic auth to it.
This is useful for dev environments where you want to share a static site with your team but ensure that it’s not publicly accessible.
This works by injecting some code into a CloudFront function that checks the basic auth
header and matches it against the USERNAME
and PASSWORD
secrets.
To deploy this, you need to first set the USERNAME
and PASSWORD
secrets.
If you are deploying this to preview environments, you might want to set the secrets using
the --fallback
flag.
View the full example.
AWS static site
Deploy a simple HTML file as a static site with S3 and CloudFront. The website is stored in
the site/
directory.
View the full example.
AWS SvelteKit container with Redis
Creates a hit counter app with SvelteKit and Redis.
This deploys SvelteKit as a Fargate service to ECS and it’s linked to Redis.
Since our Redis cluster is in a VPC, we’ll need a tunnel to connect to it from our local machine.
This needs sudo to create a network interface on your machine. You’ll only need to do this once on your machine.
To start your app locally run.
Now if you go to http://localhost:5173
you’ll see a counter update as you refresh the page.
Finally, you can deploy it by adding the Dockerfile
that’s included in this example and
running npx sst deploy --stage production
.
View the full example.
Swift in Lambda
Deploys a simple Swift application to Lambda using the al2023
runtime.
Check out the README in the repo for more details.
View the full example.
T3 Stack in AWS
Deploy T3 stack with Drizzle and Postgres to AWS.
This example was created using create-t3-app
and the following options: tRPC, Drizzle,
no auth, Tailwind, Postgres, and the App Router.
Instead of a local database, we’ll be using an RDS Postgres database.
Similarly, for Drizzle Kit.
In our Next.js app we can access our Postgres database because we link them
both. We don’t need to use our .env
files.
To run this in dev mode run:
It’ll take a few minutes to deploy the database and the VPC.
This also starts a tunnel to let your local machine connect to the RDS Postgres database. Make sure you have it installed, you only need to do this once for your local machine.
Now in a new terminal you can run the database migrations.
We also have the Drizzle Studio start automatically in dev mode under the Studio tab.
And to make sure our credentials are available, we update our package.json
with the sst shell
CLI.
So running npm run db:push
will run Drizzle Kit with the right credentials.
To deploy this to production run:
Then run the migrations.
If you are running this locally, you’ll need to have a tunnel running.
If you are doing this in a CI/CD pipeline, you’d want your build containers to be in the same VPC.
View the full example.
Subscribe to topics
Create an SNS topic, publish to it from a function, and subscribe to it with a function and a queue.
View the full example.
Vector search
Store and search for vector data using the Vector component. Includes a seeder API that uses an LLM to generate embeddings for some movies and optionally their posters.
Once seeded, you can call the search API to query the vector database.
View the full example.
React SPA with Vite
Deploy a React single-page app (SPA) with Vite to S3 and CloudFront.
View the full example.
Cloudflare Cron
This example creates a Cloudflare Worker that runs on a schedule.
View the full example.
Cloudflare KV
This example creates a Cloudflare KV namespace and links it to a worker. Now you can use the SDK to interact with the KV namespace in your worker.
View the full example.
Link multiple secrets
You might have multiple secrets that need to be used across your app. It can be tedious to create a new secret and link it to each function or resource.
A common pattern to addresses this is to create an object with all your secrets and then link them all at once. Now when you have a new secret, you can add it to the object and it will be automatically available to all your resources.
View the full example.
Default function props
Set default props for all the functions in your app using the global $transform
.
View the full example.
Vercel domains
Creates a router that uses domains purchased through and hosted in your Vercel account.
Ensure the VERCEL_API_TOKEN
and VERCEL_TEAM_ID
environment variables are set.
View the full example.