Docker deployment workflow with git
What is the best way to deploy a docker container to a production environment?
- Add a
Dockerfileto the git repository and run
docker buildon the production system
- Commit changes to a container with
docker commitand push it to a private Docker repository and then pull it with
docker pullto the production system.
Should I run
docker commit even if I don’t change the infrastructure but just the app code?
I hope my questions are clear.
2 Solutions collect form web for “Docker deployment workflow with git”
Ideally you would have some sort of registry server and your docker containers would be up there, your production environment would pull these in and use them. You shouldn’t have to update the docker container when your app code changes, ADD /project in your Dockerfile and share it with
--volumes-from to other containers. You app should independent from the containers altogether.
There are tools out there now like fig, which let you start up a development environment with docker containers. You would then take this further and deploy your app containers in a CoreOS cluster.
Here is what I have for my productAPI, this just ADDs the project code into the container.
You shouldn’t have to change the Dockerfile all that much unless your system dependencies change for the project.
FROM phusion/baseimage MAINTAINER Alex Goretoy <email@example.com> ENV DEBIAN_FRONTEND noninteractive ENV PRODUCT_API_PATH /opt/product_api RUN mkdir -p $PRODUCT_API_PATH/ ADD . $PRODUCT_API_PATH/ RUN $PRODUCT_API_PATH/setup.sh EXPOSE 8080 CMD python $PRODUCT_API_PATH/manage.py runserver
#!/bin/bash apt-get update apt-get install -y git \ wget \ openssl \ libssl-dev \ libffi-dev \ python-pip \ python2.7-dev \ postgresql-9.3 \ libpq-dev apt-get update apt-get install -y libcurl4-openssl-dev apt-get update wget https://bootstrap.pypa.io/ez_setup.py -O - | python pip install -r $PRODUCT_API_PATH/requirements.txt # Clean up APT when done. apt-get clean && rm -rf /var/lib/apt/lists/* /tmp/* /var/tmp/*
I would say that in a production environment you just want to pull and run the last image(s) of your container(s).
So the idea of having a private registry is good, and your delivery pipeline would be :
- Commit changes to your app or your Dockerfile(s)
- CI takes changes, compile your app, build your new image(s), and push it to the registry
- Pull new image(s) on the production
- Run it !
On my side, I don’t use
docker commit, i prefer tags with
docker build ... -t .... I only use
commit when i debug a container with an interactive shell.