Linux containers and Docker are radically changing the way that applications are developed, built, distributed and deployed. The AtoM team is experimenting with new workflows that make use of containers. This document introduces our new development workflow based on Docker and Docker Compose. The latter is a tool that help us to run multi-container applications like AtoM and it is suitable for both development and production environments.
The Docker Compose integration and AtoM Dockerfile outlined here are intended for testing and development purposes only. They have not been fully tested and prepared for use in production environments at this time.
If you encounter issues or bugs, you are welcome to report them via the AtoM users forum - we will provide basic support as we are able, but urge you to proceed at your own risk.
If this is the first time that you have heard about containers you may find The New Stack eBook Series a useful resource to get up to speed quickly.
Install Docker and Docker Compose¶
Docker works the same whether you are using Docker for Windows, Docker for Mac or Docker on Linux. For the latter, make sure that both Docker Engine and Docker Compose are installed following the instructions in the links.
Spin it up¶
Let’s make sure that the Docker client can reach the engine. The following command will list the currently running containers:
Now using git, check out the sources of AtoM and change your current directory:
git clone -b qa/2.6.x https://github.com/artefactual/atom.git atom cd atom
The source code (the entire “atom” folder) is copied to the Docker container on the build process and used to create the instance. For example, clone the “stable/2.5.x” branch if you want to test the latest stable version of AtoM.
Now set the environment variable
COMPOSE_FILE to tell Compose what is the
location of our YAML file. You could do the same using the
-f flag but we
don’t want to do so each time we invoke the
# For bash users (most of you) export COMPOSE_FILE="$PWD/docker/docker-compose.dev.yml" # For fish users set -lx COMPOSE_FILE (pwd)/docker/docker-compose.dev.yml
It’s time to use Docker Compose in order to provision our containers:
In some operating systems, the default virtual memory limits can be low. Elasticsearch increases one of those limits when installed via packages and in the container that will be created below. However, the same limit needs to be increased in the host running this environment. Check the Elasticsearch documentation for more information.
The containers for the application use
php:7.2-fpm-alpine as their base
image. If an old version of this image has been already downloaded by the
Docker engine in the host, run
docker pull php:7.2-fpm-alpine to get the
latest version before creating the containers. It has to be based on Alpine
v3.8 or higher to be able to install some packages.
# Create and start containers. This may take a while the first time you run # it because all the images have to be downloaded (e.g. percona, memcached) # and the AtoM image has to be built. docker-compose up -d # Execute a command in the running container atom: purge database docker-compose exec atom php symfony tools:purge --demo
While you wait, take the opportunity to check out our Dockerfile, which describes the steps that are taken to build the AtoM image. It is based on Alpine Linux + PHP 7.2 and the rest of dependencies. In addition, our docker-compose.dev.yml file shows how AtoM is orchestrated together with its service dependencies. It is an environment meant to be used by developers.
That’s it! You have started the containers and put them in the background, populated the database and initiated the Elasticsearch index. You can start developing right away. Changes made in the source code will take effect immediately.
Due to a bug that has not been solved yet, the AtoM worker needs to be restarted after the database is populated for the first time:
docker-compose restart atom_worker
To avoid reseting the configuration files to their default version each time the containers are created, the following files are only generated if they don’t exist already in the source code:
Docker Compose lets you perform many different actions. Please refer to the documentation for more help. For example, you can monitor the output of some of your containers as follows:
docker-compose logs -f atom atom_worker nginx
You can also scale the AtoM worker as needed:
docker-compose up -d --scale atom_worker=2
Let’s verify that two workers have subscribed to Gearman:
# Establish a TCP connection to gearmand, port 4730 docker-compose exec atom bash -c "nc gearmand 4730" # Send STATUS command STATUS 0a2a58137e05032d1140fdbd0d6dccbb-arInheritRightsJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arFileImportJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arInformationObjectXmlExportJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arActorXmlExportJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arCalculateDescendantDatesJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arXmlExportSingleFileJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arUpdatePublicationStatusJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arObjectMoveJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arInformationObjectCsvExportJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arUpdateEsIoDocumentsJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arActorCsvExportJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arRepositoryCsvExportJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arFindingAidJob 0 0 2 0a2a58137e05032d1140fdbd0d6dccbb-arGenerateReportJob 0 0 2
You could temporarily stop all the services with
docker-compose stop (which
docker-compose up -d later to start the services again) or both
stop and remove related containers, network and volumes by running:
docker-compose down --volumes
Connect to AtoM¶
You can run the following command to check the status and other information about the containers:
$ docker-compose ps Name Command State Ports ----------------------------------------------------------------------------------------------------- docker_atom_1 /atom/src/docker/entrypoin ... Up 9000/tcp docker_atom_worker_1 /atom/src/docker/entrypoin ... Up 9000/tcp docker_nginx_1 nginx -g daemon off; Up 0.0.0.0:63001->80/tcp docker_elasticsearch_1 /bin/bash bin/es-docker Up 127.0.0.1:63002->9200/tcp, 9300/tcp docker_percona_1 /docker-entrypoint.sh mysqld Up 127.0.0.1:63003->3306/tcp docker_memcached_1 docker-entrypoint.sh -p 11 ... Up 127.0.0.1:63004->11211/tcp docker_gearmand_1 docker-entrypoint.sh gearmand Up 127.0.0.1:63005->4730/tcp
As you can see in the right column:
- AtoM and its worker share port 9000, only in the containers network.
- Nginx is accessible in port 63001 from the host and from outside (if the host allows it).
- Elasticsearch can be accessed through port 63002, only from the host.
- Percona (MySQL) can be accessed through port 63003, only from the host.
- Memcached can be accessed through port 63004, only from the host.
- The Gearman server can be accessed through port 63005, only from the host.
AtoM should be now accessible from your browser. To connect from the host, use the following address: http://localhost:63001.
The default login details are: