Home > Article > Backend Development > How To: Containerize a Django and Postgres App with Docker
This article was originally published on the Shipyard Blog.
As you’re building out your Django and PostgreSQL app, you’re probably thinking about a few boxes you’d like it to check:
Using Docker and Docker Compose can help get your app ready for every stage of the development life cycle, from local to production. In this post, we’ll be covering some of the basics for customizing a Dockerfile and Compose file for a Django app with a Postgres database.
TLDR: Shipyard maintains a Django / Postgres starter application, set up to build and run with Docker and Docker Compose. Fork it here. You can use it as a project template or as a reference for your existing app.
Django is an open source Python-based web framework. It’s primarily used as a backend for web apps. Django follows the “batteries included” philosophy — it comes complete with routing support, a login framework, a web server, database tools, and much more. Django is often compared to Flask, and scores more favorably on almost all fronts.
You’re using Django apps every day. Spotify, Doordash, Instagram, Eventbrite, and Pinterest all have Django in their stacks — which speaks volumes to how extensible and scalable it can be.
Running a Django app with Docker containers unlocks several new use cases. Right off the bat, it’s an improvement to your local development workflows — making setup cleaner and more straightforward.
If you want to cloud-host your project, you’ll typically need it containerized. The great thing about Docker containers is that they can be used throughout every stage of development, from local to production. You can also distribute your Docker image so others can run it instantly without any installation or building.
At the very least, if you’re including a Dockerfile with your project, you can ensure it builds and runs identically every single time, on every system.
Our Python app needs a package manager to track, version, and install its dependencies. This helps us manage dependency inits/updates, instead of executing them individually, and preserve package versions across machines.
Both Pip and Poetry are popular dependency managers for Python, although there are quite a few others circulating around (e.g. uv, Conda, Rye).
Pip is incredibly straightforward. Users can list their packages in a requirements.txt file with their respective versions, and run pip install to set them up. Users can capture existing dependencies and their versions by running pip freeze > requirements.txt in a project’s root.
Poetry a highly-capable package manager for apps of any scale, but it’s slightly less straightforward to configure than Pip (it uses a TOML file with tables, metadata, and scripts). Poetry also uses a lockfile (poetry.lock) to “lock” dependencies at their current versions (and their dependencies by version). This way, if your project works at a certain point in time on a particular machine, this state will be preserved. Running poetry init prompts users with a series of options to generate a pyproject.toml file.
To Dockerize your Django app, you'll follow the classic Dockerfile structure (set the base image, set the working directory, etc.) and then modify it with the project-specific installation instructions, likely found in the README.
We can choose a lightweight Python image to act as our base for this Dockerfile. To browse versions by tag, check out the Python page on Docker Hub. I chose Alpine here since it’ll keep our image small:
FROM python:3.8.8-alpine3.13
Here, we will define the working directory within the Docker container. All paths mentioned after will be relative to this.
WORKDIR /srv
There are a few libraries we need to add before we can get Poetry configured:
RUN apk add --update --no-cache \ gcc \ libc-dev \ libffi-dev \ openssl-dev \ bash \ git \ libtool \ m4 \ g++ \ autoconf \ automake \ build-base \ postgresql-dev
Next, we’ll ensure we’re using the latest version of Pip, and then use that to install Poetry inside our container:
RUN pip install --upgrade pip RUN pip install poetry
This app’s dependencies are defined in our pyproject.toml and poetry.lock files. Let’s bring them over to the container’s working directory, and then install from their declarations:
ADD pyproject.toml poetry.lock ./ RUN poetry install
Now, we’ll copy over the rest of the project, and install the Django project itself within the container:
ADD src ./src RUN poetry install
Finally, we’ll run our project’s start command. In this particular app, it’ll be the command that uses Poetry to start the Django development server:
CMD ["poetry", "run", "python", "src/manage.py", "runserver", "0:8080"]
When we combine the snippets from above, we’ll get this Dockerfile:
FROM python:3.8.8-alpine3.13 WORKDIR /srv RUN apk add --update --no-cache \ gcc \ libc-dev \ libffi-dev \ openssl-dev \ bash \ git \ libtool \ m4 \ g++ \ autoconf \ automake \ build-base \ postgresql-dev RUN pip install --upgrade pip RUN pip install poetry ADD pyproject.toml poetry.lock ./ RUN poetry install ADD src ./src RUN poetry install CMD ["poetry", "run", "python", "src/manage.py", "runserver", "0:8080"]
We’re going to split this app into two services: django and postgres. Our django service will be built from our Dockerfile, containing all of our app’s local files.
For this app, we want to build the django service from our single Dockerfile and use the entire root directory as our build context path. We can set our build label accordingly:
django: build: .
We can map port 8080 on our host machine to 8080 within the container. This will also be the port we use to access our Django app — which will soon be live at http://localhost:8080.
ports: - '8080:8080'
Since our Django app is connecting to a database, we want to instruct Compose to spin up our database container (postgres) first. We’ll use the depends_on label to make sure that service is ready and running before our django service starts:
depends_on: - postgres
Since we’ll be sharing files between our host and this container, we can define a bind mount by using the volumes label. To set the volume, we’ll provide a local path, followed by a colon, followed by a path within the container. The ro flag gives the container read-only permissions for these files:
volumes: - './src:/srv/src:ro'
Combining all the options/configurations from above, our django service should look like this:
django: build: . ports: - '8080:8080' depends_on: - postgres volumes: - './src:/srv/src:ro'
Our Django app is configured to connect to a PostgreSQL database. This is defined in our settings.py:
DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql', 'NAME': 'app', 'USER': 'obscure-user', 'PASSWORD': 'obscure-password', 'HOST': 'postgres', 'PORT': 5432, } }
We can migrate our existing database to its own Docker container to isolate it from the base Django app. First, let’s define a postgres service in our Compose file and pull the latest lightweight Postgres image from Docker Hub:
postgres: image: 'postgres:14.13-alpine3.20'
To configure our PostgreSQL database, we can pass in a few environment variables to set credentials and paths. You may want to consider using a Secrets Manager for this.
environment: - POSTGRES_DB=app - POSTGRES_USER=obscure-user - POSTGRES_PASSWORD=obscure-password - PGDATA=/var/lib/postgresql/data/pgdata
We can expose our container port by setting it to the default Postgres port: 5432. For this service, we’re only specifying a single port, which means that the host port will be randomized. This avoids collisions if you’re running multiple Postgres instances.
ports: - '5432'
In our postgres definition, we can add a named volume. This is different from the bind mount that we created for the django service. This will persist our data after the Postgres container spins down.
volumes: - 'postgres:/var/lib/postgresql/data'
Outside of the service definitions and at the bottom of the Compose file, we’ll declare the named postgres volume again. By doing so, we can reference it from our other services if needed.
volumes: postgres:
And here’s the resulting PostgreSQL definition in our Compose file:
postgres: image: 'postgres:14.13-alpine3.20' environment: - POSTGRES_DB=app - POSTGRES_USER=obscure-user - POSTGRES_PASSWORD=obscure-password - PGDATA=/var/lib/postgresql/data/pgdata ports: - '5432' volumes: - 'postgres:/var/lib/postgresql/data' volumes: postgres:
We can get our app production-ready by deploying it in a Shipyard application — this means we’ll get an ephemeral environment for our base branch, as well as environments for every PR we open.
Shipyard transpiles Compose files to Kubernetes manifests, so we’ll add some labels to make it Kubernetes-compatible.
Under our django service, we can add two custom Shipyard labels:
labels: shipyard.init: 'poetry run python src/manage.py migrate' shipyard.route: '/'
Next, you can go to your Shipyard dashboard. If you haven’t already, sign up for a 30-day free trial.
Click the + Application button, then select your repo, services, and import your env vars.
Once it finishes building, you can click the green Visit button to access your short-lived ephemeral environment. What comes next?
Now you have a fully containerized Django app with a database! You can run it locally with Docker Compose, preview and test it in an ephemeral environment, and iterate until it’s production-ready.
Want to Dockerize a Yarn app next? Check out our guide!
The above is the detailed content of How To: Containerize a Django and Postgres App with Docker. For more information, please follow other related articles on the PHP Chinese website!