Just a heads up: On March 24, 2025, starting at 4:30pm CDT / 19:30 UTC, the site will be undergoing scheduled maintenance for a few hours. During this time, the site might be unavailable for a short while. Thanks for your patience.
×Hi
I am using in my application the package -> aws-sdk
when I am importing S3 object from the package, the docker build in bitbucket is falling with the error SIGKILL
and I don't know why its happening, on my local machine the docker build its complete fine
Hi @naorsa and welcome to the community!
I've seen this error occur when the memory available to the docker service is not enough, so that is a possible cause.
To use docker commands in your pipelines build, you define a docker service in your bitbucket-pipelines.yml file. Regular steps have 4096 MB of memory in total and each service has a 1024 MB default memory. If there are no other services in that step, you can increase that up to 3072 MB if you add in your yml file a definition as follows:
definitions:
services:
docker:
memory: 2048
Our documentation for reference:
If a step has size: 2x, the build container has 8192 MB memory in total and services can be configured to use up to 7128 MB memory (in total).
Please feel free to let me know if that helps or if you need further assistance.
Kind regards,
Theodora
Faced the same problem. Increasing docker memory to 2048 fixed it.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.
Hey Theodora,
thanks you very much, its help and the problem solved !
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.
Show up and give back by attending an Atlassian Community Event: we’ll donate $10 for every event attendee in March!
Join an Atlassian Community Event!
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.