Output k6 results to AWS Timestream so that you can run a performant, low-cost load test.
If you're here you've probably chosen to use k6 already and you're probably interested in using an AWS serverless service. These give you the benefits of:
- Performance at scale
- Low cost
- Great developer experience
For more information see the alternatives.
Using this extension lets you hook up K6 to AWS Timestream - plus you get a nice looking Grafana dashboard 😉 based off the K6 Load Testing Results dashboard.
This output is written as an extension to K6 using xk6 extensions.
You can use this extension by either:
- Taking the k6 executable from the latest release and following the instructions on running k6.
- Using the Docker image from the latest release and following the instructions on running k6.
- Building this extension into K6 - see the custom build instructions.
Include the argument --out timestream
when using the k6 run
command - see the K6 docs
For all configuration specific to this extension see the Config struct
in config.go.
The key bits of config you'll need to setup are the following environment variables
K6_TIMESTREAM_DATABASE_NAME
K6_TIMESTREAM_TABLE_NAME
You'll also need to setup your AWS credentials - see the guide on how to do this.
The timestream record dimensions (see timestream concepts) for each metric emmitted by k6 are taken from any k6 tags that have non-empty values.
Every timestream record requires at least one dimension when written, and k6 applies some default tags to metrics emmitted by many core k6 JavaScript API objects such as http requests, groups and checks. However, since some metrics emitted in the global/test scope may not have any k6 default tags, you will likely see the error At least one dimension is required for a record.
logged from timestream if you do not define at least one custom tag at the topmost scope of your script to cover metrics with no default tags, as in an options object export. More information can be found in the K6 documentation or an example of setting up tags can be found in the integration test script.
An example dashboard is provided. You can use this dashboard by running make grafana-build grafana-run
. If you are using this with your own test scripts, ensure that you include the instance_id
and vu
tags in your test script - see the integration test script as an example.
I use VSCode for development so this will be the best supported editor. However, you should be able to use other IDEs. If you are using another IDE:
- The devcontainer Dockerfile
ci
target shows all the tools you need for a dev environment (e.g. For linting). - There are suggested tools you can also use.
The preferred way to develop using VSCode is to use the dev container feature. This will mean you have all the tools required and suggested for development.
If you do want to use different tools (e.g. you don't like the shell setup), create .devcontainer/tools.override.sh
and base it off .devcontainer/tools.default.sh.
If you don't want to use dev containers, you'll need to make sure you install the tools from the devcontainer Dockerfile and the packages in suggested tools that are needed for the VSCode extensions.
output.go contains the logic for converting from K6 metric samples to AWS Timestream records and then saving those records.
There are targets for different development tasks in the Makefile.
Metric samples are passed from each of the K6 VUs to metricSamplesHandler
. This converts them to the format that the Timestream SDK expects and holds on to them until it has 100 records to save (the max batch size for Timestream). It will then save these asyncronously by kicking off a new go-routine to perform the save.
The channel for receiving metric samples is closed at the end of the test and the left-over records are saved.
graph TD;
K6-VU1.AddMetricSamples--metric samples-->metricSamplesHandler
K6-VU2.AddMetricSamples--metric samples-->metricSamplesHandler
K6-VUN.AddMetricSamples--metric samples-->metricSamplesHandler
metricSamplesHandler--have 100 samples?-->writeRecordsAsync
metricSamplesHandler--shutting down?-->writeRecordsAsync
writeRecordsAsync--new go routine-->writeRecords
The integration tests work by creating a Timestream database and table, running a load test (with a built in test script) and then checking the results.
graph LR;
Client--deploy-->Timestream;
Client--build-->k6;
Client--run-->k6;
k6-->nginx-fake-api
k6--write-->Timestream;
Client--build-->Tests;
Client--run-->Tests;
Tests--query-->Timestream;
Client--destroy-->Timestream;
To run the integration tests you'll need to setup AWS credentials - see the guide on how to do this.
To deploy the Timestream database run make deploy-infra
.
To run the tests (build, run and query steps above) run make test-integration
. Note that you will need to build the k6 image first with make build-image
.
To destroy the Timestream database run make destroy-infra
.
Testing of the Grafana dashboard is manual:
export K6_ITERATIONS=40000
- to get a reasonable number of results, set the number of iterations to a large number.make deploy-infra
- to deploy the infrastructure.make test-integration
- to run the tests. These will likely fail as the number of iterations is not what the tests expect.make grafana-build grafana-run
and browse to http://localhost:3000. From the dashboard you'll see the results come in. It should look like the dashboard near the top.make destroy-infra
- to destroy the infrastructure once you're done testing.