Security News
Supply Chain Attack Detected in Solana's web3.js Library
A supply chain attack has been detected in versions 1.95.6 and 1.95.7 of the popular @solana/web3.js library.
@aws-cdk/aws-codebuild
Advanced tools
AWS CDK v1 has reached End-of-Support on 2023-06-01. This package is no longer being updated, and users should migrate to AWS CDK v2.
For more information on how to migrate, see the Migrating to AWS CDK v2 guide.
AWS CodeBuild is a fully managed continuous integration service that compiles source code, runs tests, and produces software packages that are ready to deploy. With CodeBuild, you don’t need to provision, manage, and scale your own build servers. CodeBuild scales continuously and processes multiple builds concurrently, so your builds are not left waiting in a queue. You can get started quickly by using prepackaged build environments, or you can create custom build environments that use your own build tools. With CodeBuild, you are charged by the minute for the compute resources you use.
Install the module:
$ npm i @aws-cdk/aws-codebuild
Import it into your code:
import * as codebuild from '@aws-cdk/aws-codebuild';
The codebuild.Project
construct represents a build project resource. See the
reference documentation for a comprehensive list of initialization properties,
methods and attributes.
Build projects are usually associated with a source, which is specified via
the source
property which accepts a class that extends the Source
abstract base class.
The default is to have no source associated with the build project;
the buildSpec
option is required in that case.
Here's a CodeBuild project with no source which simply prints Hello, CodeBuild!
:
CodeCommitSource
Use an AWS CodeCommit repository as the source of this build:
import * as codecommit from '@aws-cdk/aws-codecommit';
const repository = new codecommit.Repository(this, 'MyRepo', { repositoryName: 'foo' });
new codebuild.Project(this, 'MyFirstCodeCommitProject', {
source: codebuild.Source.codeCommit({ repository }),
});
S3Source
Create a CodeBuild project with an S3 bucket as the source:
const bucket = new s3.Bucket(this, 'MyBucket');
new codebuild.Project(this, 'MyProject', {
source: codebuild.Source.s3({
bucket: bucket,
path: 'path/to/file.zip',
}),
});
The CodeBuild role will be granted to read just the given path from the given bucket
.
GitHubSource
and GitHubEnterpriseSource
These source types can be used to build code from a GitHub repository. Example:
const gitHubSource = codebuild.Source.gitHub({
owner: 'awslabs',
repo: 'aws-cdk',
webhook: true, // optional, default: true if `webhookFilters` were provided, false otherwise
webhookTriggersBatchBuild: true, // optional, default is false
webhookFilters: [
codebuild.FilterGroup
.inEventOf(codebuild.EventAction.PUSH)
.andBranchIs('master')
.andCommitMessageIs('the commit message'),
], // optional, by default all pushes and Pull Requests will trigger a build
});
To provide GitHub credentials, please either go to AWS CodeBuild Console to connect
or call ImportSourceCredentials
to persist your personal access token.
Example:
aws codebuild import-source-credentials --server-type GITHUB --auth-type PERSONAL_ACCESS_TOKEN --token <token_value>
BitBucketSource
This source type can be used to build code from a BitBucket repository.
const bbSource = codebuild.Source.bitBucket({
owner: 'owner',
repo: 'repo',
});
For all Git sources, you can fetch submodules while cloing git repo.
const gitHubSource = codebuild.Source.gitHub({
owner: 'awslabs',
repo: 'aws-cdk',
fetchSubmodules: true,
});
CodeBuild Projects can produce Artifacts and upload them to S3. For example:
declare const bucket: s3.Bucket;
const project = new codebuild.Project(this, 'MyProject', {
buildSpec: codebuild.BuildSpec.fromObject({
version: '0.2',
}),
artifacts: codebuild.Artifacts.s3({
bucket,
includeBuildId: false,
packageZip: true,
path: 'another/path',
identifier: 'AddArtifact1',
}),
});
If you'd prefer your buildspec to be rendered as YAML in the template,
use the fromObjectToYaml()
method instead of fromObject()
.
Because we've not set the name
property, this example will set the
overrideArtifactName
parameter, and produce an artifact named as defined in
the Buildspec file, uploaded to an S3 bucket (bucket
). The path will be
another/path
and the artifact will be a zipfile.
To add a CodeBuild Project as an Action to CodePipeline,
use the PipelineProject
class instead of Project
.
It's a simple class that doesn't allow you to specify sources
,
secondarySources
, artifacts
or secondaryArtifacts
,
as these are handled by setting input and output CodePipeline Artifact
instances on the Action,
instead of setting them on the Project.
const project = new codebuild.PipelineProject(this, 'Project', {
// properties as above...
})
For more details, see the readme of the @aws-cdk/@aws-codepipeline-actions
package.
You can save time when your project builds by using a cache. A cache can store reusable pieces of your build environment and use them across multiple builds. Your build project can use one of two types of caching: Amazon S3 or local. In general, S3 caching is a good option for small and intermediate build artifacts that are more expensive to build than to download. Local caching is a good option for large intermediate build artifacts because the cache is immediately available on the build host.
With S3 caching, the cache is stored in an S3 bucket which is available
regardless from what CodeBuild instance gets selected to run your CodeBuild job
on. When using S3 caching, you must also add in a cache
section to your
buildspec which indicates the files to be cached:
declare const myCachingBucket: s3.Bucket;
new codebuild.Project(this, 'Project', {
source: codebuild.Source.bitBucket({
owner: 'awslabs',
repo: 'aws-cdk',
}),
cache: codebuild.Cache.bucket(myCachingBucket),
// BuildSpec with a 'cache' section necessary for S3 caching. This can
// also come from 'buildspec.yml' in your source.
buildSpec: codebuild.BuildSpec.fromObject({
version: '0.2',
phases: {
build: {
commands: ['...'],
},
},
cache: {
paths: [
// The '**/*' is required to indicate all files in this directory
'/root/cachedir/**/*',
],
},
}),
});
Note that two different CodeBuild Projects using the same S3 bucket will not share their cache: each Project will get a unique file in the S3 bucket to store the cache in.
With local caching, the cache is stored on the codebuild instance itself. This is simple, cheap and fast, but CodeBuild cannot guarantee a reuse of instance and hence cannot guarantee cache hits. For example, when a build starts and caches files locally, if two subsequent builds start at the same time afterwards only one of those builds would get the cache. Three different cache modes are supported, which can be turned on individually.
LocalCacheMode.SOURCE
caches Git metadata for primary and secondary sources.LocalCacheMode.DOCKER_LAYER
caches existing Docker layers.LocalCacheMode.CUSTOM
caches directories you specify in the buildspec file.new codebuild.Project(this, 'Project', {
source: codebuild.Source.gitHubEnterprise({
httpsCloneUrl: 'https://my-github-enterprise.com/owner/repo',
}),
// Enable Docker AND custom caching
cache: codebuild.Cache.local(codebuild.LocalCacheMode.DOCKER_LAYER, codebuild.LocalCacheMode.CUSTOM),
// BuildSpec with a 'cache' section necessary for 'CUSTOM' caching. This can
// also come from 'buildspec.yml' in your source.
buildSpec: codebuild.BuildSpec.fromObject({
version: '0.2',
phases: {
build: {
commands: ['...'],
},
},
cache: {
paths: [
// The '**/*' is required to indicate all files in this directory
'/root/cachedir/**/*',
],
},
}),
});
By default, projects use a small instance with an Ubuntu 18.04 image. You
can use the environment
property to customize the build environment:
buildImage
defines the Docker image used. See Images below for
details on how to define build images.certificate
defines the location of a PEM encoded certificate to import.computeType
defines the instance type used for the build.privileged
can be set to true
to allow privileged access.environmentVariables
can be set at this level (and also at the project
level).The CodeBuild library supports both Linux and Windows images via the
LinuxBuildImage
(or LinuxArmBuildImage
), and WindowsBuildImage
classes, respectively.
You can specify one of the predefined Windows/Linux images by using one
of the constants such as WindowsBuildImage.WIN_SERVER_CORE_2019_BASE
,
WindowsBuildImage.WINDOWS_BASE_2_0
, LinuxBuildImage.STANDARD_2_0
, or
LinuxArmBuildImage.AMAZON_LINUX_2_ARM
.
Alternatively, you can specify a custom image using one of the static methods on
LinuxBuildImage
:
LinuxBuildImage.fromDockerRegistry(image[, { secretsManagerCredentials }])
to reference an image in any public or private Docker registry.LinuxBuildImage.fromEcrRepository(repo[, tag])
to reference an image available in an
ECR repository.LinuxBuildImage.fromAsset(parent, id, props)
to use an image created from a
local asset.LinuxBuildImage.fromCodeBuildImageId(id)
to reference a pre-defined, CodeBuild-provided Docker image.or one of the corresponding methods on WindowsBuildImage
:
WindowsBuildImage.fromDockerRegistry(image[, { secretsManagerCredentials }, imageType])
WindowsBuildImage.fromEcrRepository(repo[, tag, imageType])
WindowsBuildImage.fromAsset(parent, id, props, [, imageType])
or one of the corresponding methods on LinuxArmBuildImage
:
LinuxArmBuildImage.fromEcrRepository(repo[, tag])
Note that the WindowsBuildImage
version of the static methods accepts an optional parameter of type WindowsImageType
,
which can be either WindowsImageType.STANDARD
, the default, or WindowsImageType.SERVER_2019
:
declare const ecrRepository: ecr.Repository;
new codebuild.Project(this, 'Project', {
environment: {
buildImage: codebuild.WindowsBuildImage.fromEcrRepository(ecrRepository, 'v1.0', codebuild.WindowsImageType.SERVER_2019),
// optional certificate to include in the build image
certificate: {
bucket: s3.Bucket.fromBucketName(this, 'Bucket', 'my-bucket'),
objectKey: 'path/to/cert.pem',
},
},
// ...
})
The following example shows how to define an image from a Docker asset:
The following example shows how to define an image from an ECR repository:
The following example shows how to define an image from a private docker registry:
The class LinuxGpuBuildImage
contains constants for working with
AWS Deep Learning Container images:
new codebuild.Project(this, 'Project', {
environment: {
buildImage: codebuild.LinuxGpuBuildImage.DLC_TENSORFLOW_2_1_0_INFERENCE,
},
// ...
})
One complication is that the repositories for the DLC images are in
different accounts in different AWS regions.
In most cases, the CDK will handle providing the correct account for you;
in rare cases (for example, deploying to new regions)
where our information might be out of date,
you can always specify the account
(along with the repository name and tag)
explicitly using the awsDeepLearningContainersImage
method:
new codebuild.Project(this, 'Project', {
environment: {
buildImage: codebuild.LinuxGpuBuildImage.awsDeepLearningContainersImage(
'tensorflow-inference', '2.1.0-gpu-py36-cu101-ubuntu18.04', '123456789012'),
},
// ...
})
Alternatively, you can reference an image available in an ECR repository using the LinuxGpuBuildImage.fromEcrRepository(repo[, tag])
method.
CodeBuild lets you specify an S3 Bucket, CloudWatch Log Group or both to receive logs from your projects.
By default, logs will go to cloudwatch.
new codebuild.Project(this, 'Project', {
logging: {
cloudWatch: {
logGroup: new logs.LogGroup(this, `MyLogGroup`),
}
},
})
new codebuild.Project(this, 'Project', {
logging: {
s3: {
bucket: new s3.Bucket(this, `LogBucket`)
}
},
})
CodeBuild allows you to store credentials used when communicating with various sources, like GitHub:
new codebuild.GitHubSourceCredentials(this, 'CodeBuildGitHubCreds', {
accessToken: SecretValue.secretsManager('my-token'),
});
// GitHub Enterprise is almost the same,
// except the class is called GitHubEnterpriseSourceCredentials
and BitBucket:
new codebuild.BitBucketSourceCredentials(this, 'CodeBuildBitBucketCreds', {
username: SecretValue.secretsManager('my-bitbucket-creds', { jsonField: 'username' }),
password: SecretValue.secretsManager('my-bitbucket-creds', { jsonField: 'password' }),
});
Note: the credentials are global to a given account in a given region -
they are not defined per CodeBuild project.
CodeBuild only allows storing a single credential of a given type
(GitHub, GitHub Enterprise or BitBucket)
in a given account in a given region -
any attempt to save more than one will result in an error.
You can use the list-source-credentials
AWS CLI operation
to inspect what credentials are stored in your account.
You can specify a test report in your buildspec:
const project = new codebuild.Project(this, 'Project', {
buildSpec: codebuild.BuildSpec.fromObject({
// ...
reports: {
myReport: {
files: '**/*',
'base-directory': 'build/test-results',
},
},
}),
});
This will create a new test report group,
with the name <ProjectName>-myReport
.
The project's role in the CDK will always be granted permissions to create and use report groups with names starting with the project's name; if you'd rather not have those permissions added, you can opt out of it when creating the project:
declare const source: codebuild.Source;
const project = new codebuild.Project(this, 'Project', {
source,
grantReportGroupPermissions: false,
});
Alternatively, you can specify an ARN of an existing resource group, instead of a simple name, in your buildspec:
declare const source: codebuild.Source;
// create a new ReportGroup
const reportGroup = new codebuild.ReportGroup(this, 'ReportGroup');
const project = new codebuild.Project(this, 'Project', {
source,
buildSpec: codebuild.BuildSpec.fromObject({
// ...
reports: {
[reportGroup.reportGroupArn]: {
files: '**/*',
'base-directory': 'build/test-results',
},
},
}),
});
If you do that, you need to grant the project's role permissions to write reports to that report group:
declare const project: codebuild.Project;
declare const reportGroup: codebuild.ReportGroup;
reportGroup.grantWrite(project);
For more information on the test reports feature, see the AWS CodeBuild documentation.
CodeBuild projects can be used either as a source for events or be triggered by events via an event rule.
The @aws-cdk/aws-events-targets.CodeBuildProject
allows using an AWS CodeBuild
project as a AWS CloudWatch event rule target:
// start build when a commit is pushed
import * as codecommit from '@aws-cdk/aws-codecommit';
import * as targets from '@aws-cdk/aws-events-targets';
declare const codeCommitRepository: codecommit.Repository;
declare const project: codebuild.Project;
codeCommitRepository.onCommit('OnCommit', {
target: new targets.CodeBuildProject(project),
});
To define Amazon CloudWatch event rules for build projects, use one of the onXxx
methods:
import * as targets from '@aws-cdk/aws-events-targets';
declare const fn: lambda.Function;
declare const project: codebuild.Project;
const rule = project.onStateChange('BuildStateChange', {
target: new targets.LambdaFunction(fn)
});
To define CodeStar Notification rules for Projects, use one of the notifyOnXxx()
methods.
They are very similar to onXxx()
methods for CloudWatch events:
import * as chatbot from '@aws-cdk/aws-chatbot';
declare const project: codebuild.Project;
const target = new chatbot.SlackChannelConfiguration(this, 'MySlackChannel', {
slackChannelConfigurationName: 'YOUR_CHANNEL_NAME',
slackWorkspaceId: 'YOUR_SLACK_WORKSPACE_ID',
slackChannelId: 'YOUR_SLACK_CHANNEL_ID',
});
const rule = project.notifyOnBuildSucceeded('NotifyOnBuildSucceeded', target);
CodeBuild Projects can get their sources from multiple places, and produce multiple outputs. For example:
import * as codecommit from '@aws-cdk/aws-codecommit';
declare const repo: codecommit.Repository;
declare const bucket: s3.Bucket;
const project = new codebuild.Project(this, 'MyProject', {
secondarySources: [
codebuild.Source.codeCommit({
identifier: 'source2',
repository: repo,
}),
],
secondaryArtifacts: [
codebuild.Artifacts.s3({
identifier: 'artifact2',
bucket: bucket,
path: 'some/path',
name: 'file.zip',
}),
],
// ...
});
Note that the identifier
property is required for both secondary sources and
artifacts.
The contents of the secondary source is available to the build under the
directory specified by the CODEBUILD_SRC_DIR_<identifier>
environment variable
(so, CODEBUILD_SRC_DIR_source2
in the above case).
The secondary artifacts have their own section in the buildspec, under the
regular artifacts
one. Each secondary artifact has its own section, beginning
with their identifier.
So, a buildspec for the above Project could look something like this:
const project = new codebuild.Project(this, 'MyProject', {
// secondary sources and artifacts as above...
buildSpec: codebuild.BuildSpec.fromObject({
version: '0.2',
phases: {
build: {
commands: [
'cd $CODEBUILD_SRC_DIR_source2',
'touch output2.txt',
],
},
},
artifacts: {
'secondary-artifacts': {
'artifact2': {
'base-directory': '$CODEBUILD_SRC_DIR_source2',
'files': [
'output2.txt',
],
},
},
},
}),
});
Typically, resources in an VPC are not accessible by AWS CodeBuild. To enable access, you must provide additional VPC-specific configuration information as part of your CodeBuild project configuration. This includes the VPC ID, the VPC subnet IDs, and the VPC security group IDs. VPC-enabled builds are then able to access resources inside your VPC.
For further Information see https://docs.aws.amazon.com/codebuild/latest/userguide/vpc-support.html
Use Cases VPC connectivity from AWS CodeBuild builds makes it possible to:
Your builds can access any resource that's hosted in your VPC.
Enable Amazon VPC Access in your CodeBuild Projects
Pass the VPC when defining your Project, then make sure to
give the CodeBuild's security group the right permissions
to access the resources that it needs by using the
connections
object.
For example:
declare const loadBalancer: elbv2.ApplicationLoadBalancer;
const vpc = new ec2.Vpc(this, 'MyVPC');
const project = new codebuild.Project(this, 'MyProject', {
vpc: vpc,
buildSpec: codebuild.BuildSpec.fromObject({
// ...
}),
});
project.connections.allowTo(loadBalancer, ec2.Port.tcp(443));
Add support for CodeBuild to build on AWS EFS file system mounts using
the new ProjectFileSystemLocation.
The fileSystemLocations
property which accepts a list ProjectFileSystemLocation
as represented by the interface IFileSystemLocations
.
The only supported file system type is EFS
.
For example:
new codebuild.Project(this, 'MyProject', {
buildSpec: codebuild.BuildSpec.fromObject({
version: '0.2',
}),
fileSystemLocations: [
codebuild.FileSystemLocation.efs({
identifier: "myidentifier2",
location: "myclodation.mydnsroot.com:/loc",
mountPoint: "/media",
mountOptions: "opts"
})
]
});
Here's a CodeBuild project with a simple example that creates a project mounted on AWS EFS:
To enable batch builds you should call enableBatchBuilds()
on the project instance.
It returns an object containing the batch service role that was created,
or undefined
if batch builds could not be enabled, for example if the project was imported.
declare const source: codebuild.Source;
const project = new codebuild.Project(this, 'MyProject', { source, });
if (project.enableBatchBuilds()) {
console.log('Batch builds were enabled');
}
There are two types of timeouts that can be set when creating your Project.
The timeout
property can be used to set an upper limit on how long your Project is able to run without being marked as completed.
The default is 60 minutes.
An example of overriding the default follows.
new codebuild.Project(this, 'MyProject', {
timeout: Duration.minutes(90)
});
The queuedTimeout
property can be used to set an upper limit on how your Project remains queued to run.
There is no default value for this property.
As an example, to allow your Project to queue for up to thirty (30) minutes before the build fails,
use the following code.
new codebuild.Project(this, 'MyProject', {
queuedTimeout: Duration.minutes(30)
});
By default if a new build is triggered it will be run even if there is a previous build already in progress. It is possible to limit the maximum concurrent builds to value between 1 and the account specific maximum limit. By default there is no explicit limit.
new codebuild.Project(this, 'MyProject', {
concurrentBuildLimit: 1
});
FAQs
The CDK Construct Library for AWS::CodeBuild
The npm package @aws-cdk/aws-codebuild receives a total of 33,798 weekly downloads. As such, @aws-cdk/aws-codebuild popularity was classified as popular.
We found that @aws-cdk/aws-codebuild demonstrated a not healthy version release cadence and project activity because the last version was released a year ago. It has 4 open source maintainers collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
A supply chain attack has been detected in versions 1.95.6 and 1.95.7 of the popular @solana/web3.js library.
Research
Security News
A malicious npm package targets Solana developers, rerouting funds in 2% of transactions to a hardcoded address.
Security News
Research
Socket researchers have discovered malicious npm packages targeting crypto developers, stealing credentials and wallet data using spyware delivered through typosquats of popular cryptographic libraries.