Managing pipeline secrets
This page provides guidance on best practices for managing your secrets in a hybrid Buildkite architecture with self-hosted agents in your own infrastructure, or using Buildkite hosted agents. These secrets may be required by your Buildkite pipelines to access 3rd party systems as part of your build or deployment processes. However, these best practice guidelines help ensure that your secrets stay safely within your infrastructure and are never stored in, or sent to Buildkite.
Using a secrets storage service
The best practice for managing secrets with Buildkite is to house your secrets within your own secrets storage service, such as AWS Secrets Manager or Hashicorp Vault.
Buildkite provides various plugins that integrate reading and exposing secrets to your build steps using secrets storage services, such as the following. If a plugin for the service you use is not listed below or in Buildkite's plugins directory, please contact support.
A secrets storage service can be used with either self-hosted agents in your own infrastructure, as part of a hybrid Buildkite architecture, or with Buildkite hosted agents.
Service | Plugin |
---|---|
AWS SSM | aws-assume-role-with-web-identity-buildkite-plugin |
GC Secrets | gcp-workload-identity-federation-buildkite-plugin |
Hashicorp Vault | vault-secrets-buildkite-plugin |
Without a secrets storage service
While using a secrets storage service is the best practice for managing your secrets, if you don't or cannot use such a service and you use self-hosted agents in your own infrastructure, as part of a hybrid Buildkite architecture, this section provides alternative approaches to managing your pipeline secrets.
Exporting secrets with environment hooks
You can use the Buildkite agent's environment
hook to export secrets to a job.
The environment
hook is a shell script that is sourced at the beginning of a job.
It runs within the job's shell, so you can use it to conditionally run commands and export secrets within the job.
By default, the environment
hook file is stored in the agent's hooks
directory.
The path to this directory varies by platform; read the installation instructions for the path on your platform.
The path can also be overridden by the hooks-path
setting.
For example, to expose a Test Analytics API token to a specific pipeline, create an environment
script in your agent's hooks
directory that checks for the pipeline slug before exporting the secret:
Adding conditional checks, such as the pipeline slug and step identifier, helps to limit accidental disclosure of secrets.
For example, suppose you have a step that runs a script expecting a SECRET_DEPLOYMENT_ACCESS_TOKEN
environment variable, like this one:
In your environment
hook, you can export the deployment token only when when the job is the deployment step in a specific pipeline:
The script exports SECRET_DEPLOYMENT_ACCESS_TOKEN
only for the named pipeline and step.
Since this script runs for every job, you can extend it to selectively export all of the secrets used on that agent.
Storing secrets with the Elastic CI Stack for AWS
When using the Elastic CI Stack for AWS with your own AWS account and environment, you can store your secrets inside your stack's encrypted S3 bucket.
Unlike hooks defined in agent hooks-path
,
the Elastic CI Stack for AWS's env
hooks are defined per-pipeline.
For example, to expose a GITHUB_MY_APP_DEPLOYMENT_ACCESS_TOKEN
environment
variable to a step with identifier trigger-github-deploy
, you would create the
following env
file on your local development machine:
You then upload the env
file, encrypted, into the secrets S3 bucket with the
following command:
# Upload the env
aws s3 cp --acl private --sse aws:kms env "s3://elastic-ci-stack-my-stack-secrets-bucket/env"
# Remove the original file
rm env
See the Elastic CI Stack for AWS readme for more information and examples.